I have a very simple webapp that runs within HTML5's Canvas that doesn't have any public files that need to be indexed by search engines (beyond the front-page HTML file that includes calls to all the necessary resources). As such, I don't really need robots.txt
file, since they'll just see the public files and that's it.
Now, as a joke, I'd like to return an HTTP-418 AKA "I'm a tea pot" response every time a web-crawler asks for robots.txt
. However, if this will end up screwing me over in terms of my location in search results, then this is not a joke that would be very worthwhile for me.
Does anybody know anything about how different web-crawlers will respond to non-standard (though in this case it technically is standard) HTTP codes?
Also, on a more serious note, is there any reason to have a robots.txt
file that says "everything is indexable!" instead of just not having a file?