<feed xmlns='http://www.w3.org/2005/Atom'>
<title>cgit/robots.txt, branch master</title>
<subtitle>A hyperfast web frontend for git repositories written in C.</subtitle>
<id>https://git.zx2c4.com/cgit/atom/robots.txt?h=master</id>
<link rel='self' href='https://git.zx2c4.com/cgit/atom/robots.txt?h=master'/>
<link rel='alternate' type='text/html' href='https://git.zx2c4.com/cgit/'/>
<updated>2022-12-19T14:14:05Z</updated>
<entry>
<title>ui-tree,ui-blame: bail from blame if blob is binary</title>
<updated>2022-12-19T14:14:05Z</updated>
<author>
<name>June McEnroe</name>
<email>june@causal.agency</email>
</author>
<published>2019-12-18T21:30:12Z</published>
<link rel='alternate' type='text/html' href='https://git.zx2c4.com/cgit/commit/?id=cc6d9cc7fc010db9be6c2d90fd054fb2d189d629'/>
<id>urn:sha1:cc6d9cc7fc010db9be6c2d90fd054fb2d189d629</id>
<content type='text'>
This avoids piping binary blobs through the source-filter. Also prevent
robots from crawling it, since it's expensive.

Signed-off-by: Jason A. Donenfeld &lt;Jason@zx2c4.com&gt;
</content>
</entry>
<entry>
<title>robots.txt: disallow access to snapshots</title>
<updated>2013-08-12T19:14:10Z</updated>
<author>
<name>Jason A. Donenfeld</name>
<email>Jason@zx2c4.com</email>
</author>
<published>2013-05-28T12:17:00Z</published>
<link rel='alternate' type='text/html' href='https://git.zx2c4.com/cgit/commit/?id=23debef62104c70600be2b745ec3957538eeac6e'/>
<id>urn:sha1:23debef62104c70600be2b745ec3957538eeac6e</id>
<content type='text'>
My dmesg is filled with the oom killer bringing down processes while the
Bingbot downloads every snapshot for every commit of the Linux kernel in
tar.xz format. Sure, I should be running with memory limits, and now I'm
using cgroups, but a more general solution is to prevent crawlers from
wasting resources like that in the first place.

Suggested-by: Natanael Copa &lt;ncopa@alpinelinux.org&gt;
Suggested-by: Julius Plenz &lt;plenz@cis.fu-berlin.de&gt;
Signed-off-by: Jason A. Donenfeld &lt;Jason@zx2c4.com&gt;
</content>
</entry>
</feed>
