157
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 29 Aug 2025
157 points (100.0% liked)
Open Source
40238 readers
233 users here now
All about open source! Feel free to ask questions, and share news, and interesting stuff!
Useful Links
- Open Source Initiative
- Free Software Foundation
- Electronic Frontier Foundation
- Software Freedom Conservancy
- It's FOSS
- Android FOSS Apps Megathread
Rules
- Posts must be relevant to the open source ideology
- No NSFW content
- No hate speech, bigotry, etc
Related Communities
- !libre_culture@lemmy.ml
- !libre_software@lemmy.ml
- !libre_hardware@lemmy.ml
- !linux@lemmy.ml
- !technology@lemmy.ml
Community icon from opensource.org, but we are not affiliated with them.
founded 6 years ago
MODERATORS
Is there any reason you chose vosk over whisper.c++?
I'm not familiar with whisper.c++ but I did try faster-whisper. Unfortunately, the transcriptions took upward of 40sec and it didn't offer live transcription, which is a nice feature of vosk. There's a comparison in the readme with other differences. That said, it should be relatively modular. It shouldn't take much to swap it back to whisper if that's what you prefer to use. Whisper is in the nix flake as optional, and the program allows you to change models but i haven't bothered trying to switch back to Whisper since Vosk has been more performant.
I should also add that if you don't want to reconfigure Jambi to use Whisper, you can try WhisperNow, which is built in Python and uses Whisper. I saw similar transcription performance when I used Whisper with Jambi and decided to move forward with Vosk after testing Whisper in both programs.
Probably just playing it safe since vosk has been around longer iirc Is kinda weird to rust and not choose the more performant whisper tho