summaryrefslogtreecommitdiff
path: root/robots.txt
diff options
context:
space:
mode:
authorJason A. Donenfeld <Jason@zx2c4.com>2013-05-28 12:17:00 +0000
committerJason A. Donenfeld <Jason@zx2c4.com>2013-08-12 19:14:10 +0000
commit23debef62104c70600be2b745ec3957538eeac6e (patch)
tree3b68eab7624907836b0e614328a529e686224830 /robots.txt
parent830eb6f6ff9e6dded2bfae99d7cb46f049790470 (diff)
downloadcgit-23debef62104c70600be2b745ec3957538eeac6e.tar
cgit-23debef62104c70600be2b745ec3957538eeac6e.tar.gz
cgit-23debef62104c70600be2b745ec3957538eeac6e.zip
robots.txt: disallow access to snapshots
My dmesg is filled with the oom killer bringing down processes while the Bingbot downloads every snapshot for every commit of the Linux kernel in tar.xz format. Sure, I should be running with memory limits, and now I'm using cgroups, but a more general solution is to prevent crawlers from wasting resources like that in the first place. Suggested-by: Natanael Copa <ncopa@alpinelinux.org> Suggested-by: Julius Plenz <plenz@cis.fu-berlin.de> Signed-off-by: Jason A. Donenfeld <Jason@zx2c4.com>
Diffstat (limited to 'robots.txt')
-rw-r--r--robots.txt3
1 files changed, 3 insertions, 0 deletions
diff --git a/robots.txt b/robots.txt
new file mode 100644
index 0000000..4ce948f
--- /dev/null
+++ b/robots.txt
@@ -0,0 +1,3 @@
+User-agent: *
+Disallow: /*/snapshot/*
+Allow: /