Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

A Chat endpoint #222

Closed
glennfu opened this issue Jun 9, 2023 · 1 comment · Fixed by #493
Closed

A Chat endpoint #222

glennfu opened this issue Jun 9, 2023 · 1 comment · Fixed by #493
Labels
enhancement New feature or request

Comments

@glennfu
Copy link

glennfu commented Jun 9, 2023

Please describe the feature you want

A Chat endpoint/ui for talking about your project. Feed it your project source, then have a ChatGPT-like conversation about the project.

Additional context

Since the model already has the context of your source code, only offering code completion feels like a missed opportunity. 99% of the time I want AI to help with my project, I'm feeding ChatGPT 4 the contents of a few relevant files and then asking it for its advice on a bigger solution. The answer might affect several files in one response. Some of the useful response might not even be actual code.

All other active projects I've found that even come close to this rely on cloud models and they can't even help with anything beyond boilerplate on new projects because it's so slow and expensive (and sometimes hard-limited) to feed it the context of an already established project.

Tabby solves that with local models. I'd love to see that potential tapped into for big-picture help and not just isolated code completions.

Could this be a possible future feature for Tabby?


Please reply with a 👍 if you want this feature.

@glennfu glennfu added the enhancement New feature or request label Jun 9, 2023
@wsxiaoys
Copy link
Member

wsxiaoys commented Jun 12, 2023

Hi @glennfu

Thank you for your feedback. This is actually something that has been on our minds for a long time, and it is also the main reason that drove me to switch Tabby's underlying implementation from Triton FasterTransformer to CTranslate2 (which provides a much easier way to implement new ops like MQA used in starcoder / santacoder).

We are planning to release something w/ StarCoder around mid Q-3. Please stay tuned for updates!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants