-
Notifications
You must be signed in to change notification settings - Fork 131
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
support for BatchedInferencePipeline #169
Comments
|
fedirz
pushed a commit
that referenced
this issue
Dec 17, 2024
Merged
fedirz
pushed a commit
that referenced
this issue
Dec 17, 2024
fedirz
pushed a commit
that referenced
this issue
Dec 22, 2024
fedirz
pushed a commit
that referenced
this issue
Dec 22, 2024
fedirz
pushed a commit
that referenced
this issue
Dec 22, 2024
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
faster whisper recently added BatchedInference support in version 1.1.0, allowing up to 4x faster transcription when transcribing large files.
Is there any plan to add this to this server?
Here are my test results, on my RTX 3090:
(This was for a 1 hour audio file, Total Time includes the time it takes to load the model)
I did some testing using the faster whisper server, and its speed matches up with large v2 no batch.
When using smaller audio files (30 seconds) the speed up was around 24% on batch 8 and 16.
Additionally: can we treat the requests of multiple different users as one batch?
The text was updated successfully, but these errors were encountered: