We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
https://en.wikipedia.org/wiki/Robots_exclusion_standard
Any request that isnt allowed by robots.txt should be reported as such.
The text was updated successfully, but these errors were encountered:
(participating)
Sorry, something went wrong.
We can use https://docs.python.org/2/library/robotparser.html for this, and with some enhancement like cache robots.txt for each site, etc..
URLHeadBear.py: Use robots.txt
be132e1
Requests that are not allowed by robots.txt are reported. Closes coala#1782
PrajwalM2212
Successfully merging a pull request may close this issue.
https://en.wikipedia.org/wiki/Robots_exclusion_standard
Any request that isnt allowed by robots.txt should be reported as such.
The text was updated successfully, but these errors were encountered: