The main reason for writing crawl was the lack of simple open source web crawlers. Crawl is only a few thousand lines of code and fairly easy to debug and customize.
The current version of Crawl identifies itself as Crawl/0.4 libcrawl/0.1 to web servers. It's default configuration also limits how often a fetch can happen against the same web server.
$ crawl -m 0 http://www.w3.org/