Notes
- Added audio generation endpoints, which is based on the OpenAI Speech API . It uses TikTok's Text-To-Speech to provide the voices (there's a ton to use). This is useful for those who run their own Web UI's for LLMs and would like to have their messages read to them. Or perhaps you are a developer who is making/prototyping a voice-based assistant but you do not want to pay for speech generation APIs during development.
The speech API can either be used alongside the /completions endpoint (to "hear" your LLM's responses) or it can be used independently so you can generate speech based on your own input text, again mimicking how OpenAI does it for compatibility. Refer to the website to see the sample API calls.
- /v1/audio/speech/
- /v1/audio/speech/voices
If you are interested in what voices you can use you can take a look here
Here is a demo of it in action
Screen.Recording.2024-11-20.at.2.37.22.AM.mov
These endpoints are completely optional are not functional without setting a session_id
as outlined in the Wiki
-
The partial-context bug that was supposed to ship with Version 1.0.0, is now shipped to this current version. I had made the mistake of not applying the bug fix to the branch I was working on at the time. But all is well now.
-
A Wiki has been made for setup instructions, so it declutters the main README file.
-
The website has been updated for this release, check it out