-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Request: smaller dependency surface on Transformers model types? #408
Comments
I think That said, if this is something you go after, here are a couple relevant |
In fact the branch pythonic they are developing take a long time but this is still active. Just they forget to communicate with community in their movement. For controlled generation you have LMQL, but personally I find it quite hard to use compared to a guidance. They make many stuff hiding most of what happen and making it hard to debug when things go wrong (and this is often the case, I have unexplainable memory leak with LQML). Also, you have |
@paucodeici I agree with everything you've said about lmql. It's a pain to use and debug. I still haven't found anything better yet though. Guidance was way better, but alas, it seems pretty dead. Have you tried the |
The way it will be handled seem promising when you look the commit (at
least it will be way more pythonic but i feel they are still not sure how
to expose the things to be "pythonic").
From my last test it was not working for my use case but this is a work in
progress branch so it has to be expected things can be broken.
Le mar. 7 nov. 2023 à 21:26, Josh Freckleton ***@***.***> a
écrit :
… @paucodeici <https://github.com/paucodeici> I agree with everything
you've said about lmql. It's a pain to use and debug. I still haven't found
anything better yet though. Guidance *was* way better, but alas, it seems
pretty dead.
Have you tried the pythonic branch?
—
Reply to this email directly, view it on GitHub
<#408 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/BDTBQFD4DIM4JVOYWGJHAGTYDKKNPAVCNFSM6AAAAAA6LEB63CVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTOOJZHE2TCNJZGU>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Please check out our current attempt in the new release, where we integrate llama.cpp as well. |
Hi, this is a very vague feature request :D
There are some parts of the code (in the main branch, I haven't looked at the upcoming branches that were mentioned in #395 ) that depend on some private/internal functionality of Transformers types. Mostly around the cache and logits history.
The problem with this is, it's difficult to create wrappers for other model providers, like GPTQ, AQW, llamacpp etc etc because you have to replicate some pretty specific private code from the Transformer types that Guidance expects to see.
So my request is this - is it possible to define a smaller interface/surface, not specific to Transformers (or at least not relying on anything internal to Transformers), that can be used to host models from other sources?
No idea how difficult this is or if it's realistic, just something that I would personally love to see.
The text was updated successfully, but these errors were encountered: