-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TPOT stuck at 0% #542
Comments
I think this is related to #508. Please try to run TPOT like this demo below:
Please let me know if this way solve this issue. |
@weixuanfu Thank you for the prompt reply. I should have mentioned that I am using Windows. Sorry about that. 'fork server' doesn't work on windows. How can I make it work for Windows? I have changed n_jobs =1, and even that doesn't seem to work. Thanks again! |
Oh, I just found that. Maybe decreasing n_jobs to 1 would help for Windows. Or you could try the latest dev branch where has better timeout control. Please let me know and inform me more environment infos (Tpot and its deps' versions) if both possible solutions do not work. I need double-check it. |
@weixuanfu n_jobs =1 worked for windows. I am also running it on a linux box with with n_jobs=20. It seems to be working on linux. |
I have the exact same issue, running on Windows. Even with the below params running on the tiny Titanic dataset (100's of rows), the optimizer simply never makes progress.
That said, CPU usage is around 100% and python processes are constantly getting spun up and torn down, but no progress is made. Have any of the devs tried multiprocessing on a Windows machine? I suspect it just doesn't work.
|
Multiprocessing simply doesn't work in Windows with Python, so we had to drop support for it. |
Closing this issue. Please feel free to re-open or file a new issue if you have any further questions or comments. |
I am just rewrite my code like this demo,but it still stuck at 0% for two days. |
1 similar comment
I am just rewrite my code like this demo,but it still stuck at 0% for two days. |
@OhMyGodness could you please try Parallel Training with Dask for this big dataset? |
Yes,I have done some try using dask just like this demo ,but ,but find some other mistake in the script way on my aws |
I am using TPot classifier for a smallish dataset with 20,000 rows and 68 features. I ran the following code
pipeline_optimizer = TPOTClassifier()
pipeline_optimizer = TPOTClassifier(generations=5, population_size=20, cv=5, random_state=0, verbosity=2,n_jobs = 10)
X_train = np.nan_to_num(X_train)
pipeline_optimizer.fit(X_train, dataY_train)
Warning: Although parallelization is currently supported in TPOT for Windows, pressing Ctrl+C will freeze the optimization process without saving the best pipeline! Thus, Please DO NOT press Ctrl+C during the optimization procss if n_jobs is not equal to 1. For quick test in Windows, please set n_jobs to 1 for saving the best pipeline in the middle of the optimization process via Ctrl+C.
Optimization Progress: 0%| | 0/120 [00:00<?, ?pipeline/s]
The optimization process is stuff at 0% for last 14 hours. Is this normal? Any help would be appreciated. Thank you!
The text was updated successfully, but these errors were encountered: