- 
          
- 
                Notifications
    You must be signed in to change notification settings 
- Fork 10.9k
[Bugfix] add input embedding #11684
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bugfix] add input embedding #11684
Conversation
| 👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can do one of these: 
 🚀 | 
0429875    to
    8c7751d      
    Compare
  
    | Thanks for opening this PR, can you explain what this PR is about and how this is related to #11375? | 
| 
 @DarkLight1337 Sorry for referencing the incorrect issue number. Please refer to the following issues for | 
5963444    to
    f70bbb3      
    Compare
  
    5c82b7a    to
    868a730      
    Compare
  
    | This pull request has merge conflicts that must be resolved before it can be | 
| @Bryce1010 Hi, when I used your branch, I ran the following case But I came out of the error like  | 
| @Bryce1010 I changed the  Now the error is  | 
| @Bryce1010 @DarkLight1337 Is there an expected time for when this will be updated/merged? | 
| I would love to see this feature merged. Is there a timeline for when you expect this to be reviewed and merged @njhill and @alexm-redhat ? | 
209e8ae    to
    868a730      
    Compare
  
    868a730    to
    cd8fcf8      
    Compare
  
    | 
 @DarkLight1337 Thanks for your response! The code has been rebased. Could you please review it and let me know if any further changes are needed before merging into the main branch? | 
3fdc9b5    to
    3451391      
    Compare
  
    | Can you fix the pre-commit errors? | 
6f480cb    to
    c51e4a1      
    Compare
  
    c51e4a1    to
    fa8caec      
    Compare
  
    | There are still errors, please fix them | 
| This pull request has merge conflicts that must be resolved before it can be | 
| I would like to ask when this feature will be merged into the main branch. Once merged, will it support Python 3.8? | 
| Support for Python 3.8 in general has been dropped a while ago as it has reached EOL. | 
| Would love to see this one merged! | 
| @Bryce1010 @DarkLight1337 Since it has been a few weeks without an update, and I would love to have this feature, I took the initiative to rebase this PR onto main and fix the pre-commit errors as previously asked. I can't commit to this branch, since it's owned by @Bryce1010, so these updates are in #15428. | 
| 
 That sounds good, thanks for picking this up! | 
| 
 My goat | 
| Closing in favor of #15428 | 
adds support for passing prompt_embeds to LLM.generate as
or
this enables use cases when only the embedding layer is finetuned, and have the same model backend support multiple custom tuned embedding layers
FIX #416
FIX #8323
FIX #14621