Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support reading sitemap from robots.txt #3

Open
mehaase opened this issue Jan 16, 2019 · 0 comments
Open

Support reading sitemap from robots.txt #3

mehaase opened this issue Jan 16, 2019 · 0 comments
Labels
enhancement New feature or request

Comments

@mehaase
Copy link
Contributor

mehaase commented Jan 16, 2019

We currently use robots.txt only for access control decisions. It would be nice if we could read a sitemap and add everything in it to the crawl frontier. As with other robots features, this should be controlled by a crawl policy.

@mehaase mehaase added the enhancement New feature or request label Jan 21, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant