From f17f882df292b29323f1e1da515bd16f326cee4a Mon Sep 17 00:00:00 2001 From: XhmikosR Date: Sat, 3 Aug 2019 17:09:03 +0300 Subject: [PATCH] Disable crawling on Netlify. (#29193) --- Gruntfile.js | 2 +- docs/robots.txt | 4 +++- 2 files changed, 4 insertions(+), 2 deletions(-) diff --git a/Gruntfile.js b/Gruntfile.js index 39ba49ada3..43f923fb76 100644 --- a/Gruntfile.js +++ b/Gruntfile.js @@ -291,7 +291,7 @@ module.exports = function (grunt) { docs: {}, netlify: { options: { - raw: 'github: true\nbaseurl: ""' + raw: 'github: true\nbaseurl: ""\nnetlify: true' } }, github: { diff --git a/docs/robots.txt b/docs/robots.txt index aef3fa1f64..c68014eda1 100644 --- a/docs/robots.txt +++ b/docs/robots.txt @@ -3,7 +3,9 @@ # www.robotstxt.org/ +{% if site.netlify != true and site.github == true -%} # Allow crawling of all content +{%- endif %} User-agent: * -Disallow: +Disallow:{% if site.netlify == true or site.github != true %} /{% endif %} Sitemap: {{ site.url }}/sitemap.xml