From 1ebb8e7d9b4bd52c7fc088aeaec12ab9bcb240c6 Mon Sep 17 00:00:00 2001 From: XhmikosR Date: Fri, 2 Aug 2019 17:26:25 +0300 Subject: [PATCH] robots.txt: adapt for Netlify. (#29192) Since we build with `HUGO_ENV` set to `production` on Netlify, use another variable to prevent crawling. --- site/layouts/robots.txt | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff --git a/site/layouts/robots.txt b/site/layouts/robots.txt index e9d78f1aed..c6703cd06c 100644 --- a/site/layouts/robots.txt +++ b/site/layouts/robots.txt @@ -1,8 +1,12 @@ # www.robotstxt.org -{{ if (eq (getenv "HUGO_ENV") "production") -}} +{{- $isProduction := eq (getenv "HUGO_ENV") "production" -}} +{{- $isNetlify := eq (getenv "NETLIFY") "true" -}} +{{- $allowCrawling := and (not $isNetlify) $isProduction -}} + +{{ if $allowCrawling }} # Allow crawling of all content {{- end }} User-agent: * -Disallow:{{ if (ne (getenv "HUGO_ENV") "production") }} /{{ end }} +Disallow:{{ if not $allowCrawling }} /{{ end }} Sitemap: {{ .Site.BaseURL }}/sitemap.xml