Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Website slowdown by bots #179

Open
fbastian opened this issue May 26, 2020 · 0 comments
Open

Website slowdown by bots #179

fbastian opened this issue May 26, 2020 · 0 comments

Comments

@fbastian
Copy link
Member

In GitLab by @fbastian on Feb 12, 2018, 17:19

Suggestions from Iakov about managing bots making too many requests:

  1. crawl-delay in robots.txt for yahoo & bing (if they are the problem)
  2. google webmaster tools allows you to set limit for google
  3. http code 429 + retry-after for nasty bots (if they are useless, just disallow /*)
  4. questionable: robots.txt + disallow for certain pages
  5. advanced: perhaps we discussed that already, if I were you I would use a) cache all standard elements b) save all gene pages to files (e.g. excluding header/footer). it is also possible to cache sql requests if you have enough hdd/ssd.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant