Replies: 1 comment 2 replies
-
Is this endpoint dedicate to tabby? If that's the case, you might consider deploy tabby to this machine directly? |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hey Tabby Team,
Love what you're doing with Tabby, it's a game-changer! Quick question: is there a way to route Tabby's inference requests to an external endpoint? I've got a setup that batches requests to remote GPUs for better performance.
It could lower latency and let users like me use bigger, custom models. Plus, it might be a cool feature for others too. Just a thought – totally get it if it's not doable right now.
Thanks for all the awesome work!
Beta Was this translation helpful? Give feedback.
All reactions