<feed xmlns='http://www.w3.org/2005/Atom'>
<title>cgit/robots.txt, branch lf/filter</title>
<subtitle>🍴 My patches to cgit
</subtitle>
<id>https://git.ache.one/cgit/atom?h=lf%2Ffilter</id>
<link rel='self' href='https://git.ache.one/cgit/atom?h=lf%2Ffilter'/>
<link rel='alternate' type='text/html' href='https://git.ache.one/cgit/'/>
<updated>2013-08-12T19:14:10Z</updated>
<entry>
<title>robots.txt: disallow access to snapshots</title>
<updated>2013-08-12T19:14:10Z</updated>
<author>
<name>Jason A. Donenfeld</name>
<email>Jason@zx2c4.com</email>
</author>
<published>2013-05-28T12:17:00Z</published>
<link rel='alternate' type='text/html' href='https://git.ache.one/cgit/commit/?id=23debef62104c70600be2b745ec3957538eeac6e'/>
<id>urn:sha1:23debef62104c70600be2b745ec3957538eeac6e</id>
<content type='text'>
My dmesg is filled with the oom killer bringing down processes while the
Bingbot downloads every snapshot for every commit of the Linux kernel in
tar.xz format. Sure, I should be running with memory limits, and now I'm
using cgroups, but a more general solution is to prevent crawlers from
wasting resources like that in the first place.

Suggested-by: Natanael Copa &lt;ncopa@alpinelinux.org&gt;
Suggested-by: Julius Plenz &lt;plenz@cis.fu-berlin.de&gt;
Signed-off-by: Jason A. Donenfeld &lt;Jason@zx2c4.com&gt;
</content>
</entry>
</feed>
