-
Notifications
You must be signed in to change notification settings - Fork 569
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
scrapyd not fully utilizing cpu #336
Comments
max_proc The maximum number of concurrent Scrapy process that will be started. If unset or 0 it will use the number of cpus available in the system multiplied by the value in max_proc_per_cpu option. Defaults to 0.
|
I thought from the description it means that if we leave max_proc to zero it will anyway use the maxmimum available number. But I will try as soon as the pipeline is cleared your proposal with max_proc = 2000000 |
@KarlMaresch, #173 can be a solution. Also, if you have more than one project, you are affected by #187. |
Closing as no follow-up after maintainers' advice. |
We have a lot of perodic crawler jobs to do and therefore bought a server with a strong 24 core - 48 threads amd epyc cpu. Currently there are always between 7000 & 9000 crawler jobs on pending, so there is enough work in the pipeline. However, scrapyd is only running between ~8 and ~max. 20 jobs simultaneously.
Already tried many changes on the config file, but none fixed the problem.
Current config:
The text was updated successfully, but these errors were encountered: