I got this error, first of the kind in several days of on-and-off scraping:
mechanize._response.httperror_seek_wrapper: HTTP Error 403: request disallowed by robots.txt
However, robots.txt of the site reads:
User-agent: *
Disallow:
According to this source, if the site were closed to this kind of access, robots.txt would contain Disallow: /
.
Does the error still mean that I should stop scraping, or that there is another issue?
Should I try to appease the server (like making requests less frequent), or just circumvent the error by adding headers etc.?
Finally, considering the 403, is it unethical to keep scraping?