robots.txt: disallow crawling dynamically generated PDF documents (#31617).

Patch by Go MAEDA.


git-svn-id: http://svn.redmine.org/redmine/trunk@19867 e93f8b46-1217-0410-a6f0-8f06a7374b81
This commit is contained in:
Go MAEDA
2020-07-09 00:33:41 +00:00
parent 8d2d557f2c
commit 09db3e7131
2 changed files with 4 additions and 0 deletions

View File

@@ -31,5 +31,7 @@ class WelcomeTest < Redmine::IntegrationTest
assert @response.body.match(%r{^Disallow: /projects/ecookbook/issues\r?$})
assert @response.body.match(%r{^Disallow: /issues\?sort=\r?$})
assert @response.body.match(%r{^Disallow: /issues\?\*set_filter=\r?$})
assert @response.body.match(%r{^Disallow: /issues/\*\.pdf\$\r?$})
assert @response.body.match(%r{^Disallow: /projects/\*\.pdf\$\r?$})
end
end