Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Please add lora support for higher ranks and alpha values #2847

Open
parikshitsaikia1619 opened this issue Feb 13, 2024 · 13 comments
Open

Please add lora support for higher ranks and alpha values #2847

parikshitsaikia1619 opened this issue Feb 13, 2024 · 13 comments
Labels

Comments

@parikshitsaikia1619
Copy link

ValueError: LoRA rank 64 is greater than max_lora_rank 16.

@SuperBruceJia
Copy link

Mark

@Peter-Devine
Copy link

Bump

@dspoka
Copy link

dspoka commented Mar 16, 2024

It's not super well documented but you need to just pass in "-max-lora-rank 64" or whatever when serving since default is 16.

python -m vllm.entrypoints.openai.api_server --max-lora-rank 64
--model model_name
--enable-lora
--lora-modules lora-name=lora_path

@spreadingmind
Copy link

It's not super well documented but you need to just pass in "-max-lora-rank 64" or whatever when serving since default is 16.

python -m vllm.entrypoints.openai.api_server --max-lora-rank 64 --model model_name --enable-lora --lora-modules lora-name=lora_path

Thanks for the answer, it helped me as well. For those who use code, it would be here:

          llm = LLM(
            model=args.model, tensor_parallel_size=torch.cuda.device_count(), 
            dtype=args.dtype, trust_remote_code=True, enable_lora=True, max_lora_rank=64
        )

@Napuh
Copy link

Napuh commented Apr 10, 2024

Both answers work for me, up to rank 64. Rank > 64 is not supported yet.

See #3934

@patrickrho
Copy link

patrickrho commented Jun 9, 2024

Can we get Lora rank > 64 supported and merged?

edit: im also curious if this was by design to support up to 64 rank, if so please let me know

@kevinjesse
Copy link

Bump. I need adapters that are much, much larger to be supported. Thanks

@jiangjin1999
Copy link

Is there something special about lora rank >64. Wonder why only <=64 is supported

@JohnUiterwyk
Copy link

same here, this is a blocker for me

@Peter-Devine
Copy link

Peter-Devine commented Aug 4, 2024

@JohnUiterwyk , has this not been fixed by the suggestions from @dspoka and @spreadingmind ? Their suggestions worked for me.

@JohnUiterwyk
Copy link

No, as the maximum max_lora_rank is 64, going higher than that throws an error. I have adapters with rank 128 and 256 for certain uses cases, and can not serve them with vllm as a result of the hardcoded limit for the allowed value passed to max_lora_rank

Copy link

github-actions bot commented Dec 2, 2024

This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you!

@github-actions github-actions bot added the stale label Dec 2, 2024
@AntreasAntonio
Copy link

Any updates on this? Recent papers showed that a rank=256 seems to be very beneficial for example. I suspect this trend will continue to be the case and increasing in the near future.

@github-actions github-actions bot added unstale and removed stale labels Dec 10, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests