When a scraper reads your robots.txt file don't you think they would avoid the disallowed pages and directories?
Then would you believe the scraper reads your robots.txt file a SECOND time after just downloading a few pages and immediately opens the page that it's told to leave alone and WHAMMO! gets stopped.
How FUCKING STUPID can you be to write such brain damaged code?
Sunday, March 12, 2006
Knuckle Scraping Neanderthal
Posted by IncrediBILL at 3/12/2006 01:36:00 PM
Subscribe to:
Post Comments (Atom)
4 comments:
Probably because it was a human tyring to figure out what you were hiding?
HopeSeekr of xMule
If that was the case anonymous then Bill got the title for this rant correct.
Nope, wasn't a human, no graphics ever loaded and not using Lynx.
Trust me on one thing, my profiling techniques are pretty much spot on and I test it myself several times a week to see if I could get trapped and humans can escape with a single click or two, it's not designed to hold humans.
Bots just keeping going even when being challenged to do something to prove it isn't a bot, so 20+ pages later I'm sure it's not a human.
In which case Bill your title is way off.
A better one would be "Knuckle Scraping Cylon Toasters".
Post a Comment