Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Slow/Freeze with powerful configuration #905

Open
lschneidpro opened this issue Aug 17, 2019 · 6 comments
Open

Slow/Freeze with powerful configuration #905

lschneidpro opened this issue Aug 17, 2019 · 6 comments
Labels

Comments

@lschneidpro
Copy link

lschneidpro commented Aug 17, 2019

Hi there,

With the cluster of my university, I'm trying to run TPOT for a classification problem on a jupyter notebook with a datset of size (166'158x193). I cannot provide the dataset for confidentialy reasons.

Despite:

  • powerful instance (36 cores, 180 GB RAM)
  • tuning parameters ("TPOT LIGHT", reducing generation and population_size)
  • consulting other issues (reinstall TPOT development branch,...)
  • using Dask

The fitting time is vers slow and the progress bar only progress if i stop the kernel:
Capture d’écran 2019-08-17 à 15 33 05

Here is my configuration:
Capture d’écran 2019-08-17 à 15 34 15

Here is my dask cluster:
Capture d’écran 2019-08-17 à 15 35 04

Here is my TPOT code:

pipeline_optimizer = TPOTClassifier(generations=20, 
                                    population_size=50, 
                                    scoring='neg_log_loss',
                                    cv=ut.RepeatedHoldout(n_iter=30, train_size=0.6, test_size=0.1),
                                    random_state=42, 
                                    verbosity=2,
                                    n_jobs=-1,
                                    warm_start = True,
                                    use_dask=True)
pipeline_optimizer.fit(df_cv[xvars].values, df_cv[yvars].values, weight)

where ut.RepeatedHoldout() is a personal cross-validation generator to deal with Time-dependant data
So I'm wondering if missed any important parameters to run smoothly TPOT,...

Thank you for any help

@weixuanfu
Copy link
Contributor

Did this configuration pass in smaller dataset? If yes, I think the issue maybe caused by high number of n_iter in RepeatedHoldout and max_eval_time_mins=5 in default TPOT setting. Increasing max_eval_time_mins in TPOTClassifier for allowing more time budget of evaluating a single pipeline should be helpful.

@lschneidpro
Copy link
Author

Thanks for your input.
I'll give a go.
SHould i still keep using dask ?I'm developing on a jupyter notebook

@weixuanfu
Copy link
Contributor

I think using dask should be fine.

@lschneidpro
Copy link
Author

It became worse.

I'm wondering if there is something not working with the parralel execution

Capture d’écran 2019-08-19 à 22 43 43

@weixuanfu
Copy link
Contributor

weixuanfu commented Aug 19, 2019

Did this configuration pass a test with smaller dataset? It seems the dataset is too large to finish cv even within 15 minutes time budget.

@lschneidpro
Copy link
Author

Yes, it went fine on the example provided in your documentation

Untitled.pdf

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants