Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Environment for running the code #5

Closed
lxuechen opened this issue May 24, 2023 · 9 comments
Closed

Environment for running the code #5

lxuechen opened this issue May 24, 2023 · 9 comments

Comments

@lxuechen
Copy link

Could the authors share the requirements and general environment for running this code? I am also hitting another few issues, and currently trying to infer the right versions of libraries.

@artidoro
Copy link
Owner

Hey! Thanks a lot for your interest in QLoRA. The necessary changes to use QLoRA will be merged in the transformers library tomorrow morning and we will update this repo with installation instructions.

@lxuechen
Copy link
Author

Which transformers branch / PR is this? I can install from source.

@qwopqwop200
Copy link

qwopqwop200 commented May 24, 2023

@artidoro
Copy link
Owner

FYI this is the PR: huggingface/transformers#23479

@lxuechen
Copy link
Author

FYI this is the PR: huggingface/transformers#23479

Great thanks!

@Qubitium
Copy link
Contributor

Qubitium commented May 24, 2023

@artidoro paged_adamw_32bit throws this error

  File "/root/transformers-4bit/src/transformers/utils/generic.py", line 348, in _missing_
    raise ValueError(
ValueError: paged_adamw_32bit is not a valid OptimizerNames, please select one of ['adamw_hf', 'adamw_torch', 'adamw_torch_fused', 'adamw_torch_xla', 'adamw_apex_fused', 'adafactor', 'adamw_bnb_8bit', 'adamw_anyprecision', 'sgd', 'adagrad']

Already on huggingface/transformers#23479

Env:
Cuda 12.1
Latest compiled bitsandbytes with 4bit merged.
Latest transformer PR 23479

@qwopqwop200
Copy link

@artidoro paged_adamw_32bit throws this error

  File "/root/transformers-4bit/src/transformers/utils/generic.py", line 348, in _missing_
    raise ValueError(
ValueError: paged_adamw_32bit is not a valid OptimizerNames, please select one of ['adamw_hf', 'adamw_torch', 'adamw_torch_fused', 'adamw_torch_xla', 'adamw_apex_fused', 'adafactor', 'adamw_bnb_8bit', 'adamw_anyprecision', 'sgd', 'adagrad']

Already on huggingface/transformers#23479

Try this branch.
https://github.com/2021-DGSW-Ensemble/transformers
This fork solves that problem. However, I'm running into another problem.
#3

@Qubitium
Copy link
Contributor

Qubitium commented May 24, 2023

@qwopqwop200 Confirmed https://github.com/2021-DGSW-Ensemble/transformers fixed paged_adamw_32bit not found.

@artidoro The transformer code that qlora mentions requires both the main PR 23479 and the paged lion PR to be merged to work, for now. May need to update the readme.

@artidoro
Copy link
Owner

Closing the issue as we now have info on how to install relevant packages. Let us know if you still have problems.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants