Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Ensemble bayesian learning clean #3

Open
wants to merge 16 commits into
base: reg_cocktails
Choose a base branch
from

Conversation

ravinkohli
Copy link
Owner

Types of changes

  • Breaking change (fix or feature that would cause existing functionality to not work as expected)
  • Bug fix (non-breaking change which fixes an issue)
  • New feature (non-breaking change which adds functionality)

Note that a Pull Request should only contain one of refactoring, new features or documentation changes.
Please separate these changes and send us individual PRs for each.
For more information on how to create a good pull request, please refer to The anatomy of a perfect pull request.

Checklist:

  • My code follows the code style of this project.
  • My change requires a change to the documentation.
  • I have updated the documentation accordingly.
  • Have you checked to ensure there aren't other open Pull Requests for the same update/change?
  • Have you added an explanation of what your changes do and why you'd like us to include them?
  • Have you written new tests for your core changes, as applicable?
  • Have you successfully ran tests with your changes locally?

Description

Motivation and Context

How has this been tested?

ravinkohli and others added 16 commits April 26, 2022 14:21
* add repeated kfold

* working repeated k fold

* working stacking evaluator without changing dataset and no final predict

* replace datamanager

* fix prediction with stack ensembles

* adaptive repeats

* working version of stacking with changing dataset preserving categorical info

* working version of ensemble selection per layer, TODO: send predictions according to the weights associated with the model

* finish previous todo: send predictions according to the weights associated with the model

* working version of base repeat stacked ensembles, todo: check if other methods still work, add autogluon stacking

* working all stacking versions

* rename optimisation stacking ensemble

* Add autogluon stacking  (#1)

* add working traditional models according autofluon

* working pytorch embedding with skew and embed column splitting

* work in progress: autogluon ensembling

* working autogluon ensemble

* important fix for more than 2 stacking layers

* fix for running more than 2 stacking layers

* working autogluon with default nn config from autogluon

* working xgboost model

* add configurationspace to traditional classification models

* working autogluon stacking and stacking optimisation, todo: search for autogluon models and post hoc ensemble selection for ensemble optimisation

* added post fit ensemble optimization, working per layer selection, repeat models, stacking optimisation

* update config space for search, fix stratified resampling, fix printing model with weights for soe

* fix running traditional pipeline for all the ensembles, fix get config from run history

* fix cut off num run for all ensembles

* __init__ file for column splittin

* all requirements

* add __init__.py for trad ml

* pass smbo class to custom callback

* early stop also ensemble opt

* remove -1 from autogluon stacking

* reduce number of models stored after stcking

* fix issue wiuth null identifiers in selected ensemble identifiers

* remove pointless line for debug

* set multiprocessing context to forkserver for n workers 1

* fix error when all repeats do not finish

* examples changed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant