-
-
Notifications
You must be signed in to change notification settings - Fork 235
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Integrate faster models #36
Comments
Hi @jsboige. https://github.com/guillaumekln/faster-whisper According to the repo, it reduces the VRAM usage to 12GB -> 4GB on I guess I should add an command line argument for faster-whisper and impelement it. |
That would be great ! |
faster-whisper implemented in #37, |
Thanks, I could test it with success on French songs, but language selection does not seem to work (I had to use auto-detection)
That looks underwhelming indeed. Anyway it's definitely using less VRAM, and I believe I got a perceptible performance gain for small songs' mp3s. |
Thanks for pointing this out. fixed in # 6726c6a . |
You set a default beam size of 5 for faster-whisper:
but you don't set the same beam size in openai-whisper (which is 1 by default). You should set the same beam size when comparing the transcription time. |
@guillaumekln Thanks! You're right. I didn't read it correctly. By the way, |
resolved with faster-whipser |
The large model currently brings good results, but it seems faster versions have emerged here and there, typically through quantization, with similar results, faster inference and a much lower memory footprint.
See for instance:
What would it take to support some of those?
The text was updated successfully, but these errors were encountered: