Based on our record, coqui STT should be more popular than Silero VAD. It has been mentiond 13 times since March 2021. We are tracking product recommendations and mentions on various public social media platforms and blogs. They can help you identify which product is more popular and what people think of it.
I just noticed that https://coqui.ai/ is "Shutting down". I'm building a web app (React / Django) which takes a list of affirmations & goals (in Markdown files), puts them into a database (SQlite), and uses voice synthesis to create voice audio files of the phrases. These are combined with a relaxed backing track (ffmpeg), made into playlists of 10-20 phrases (randomly sampled, or according to a theme: "mind"... - Source: Hacker News / about 3 hours ago
Not sure how relevant this is but note that Coqui TTS (the realistic TTS) has already shut down https://coqui.ai. - Source: Hacker News / about 1 month ago
You can take a look at https://coqui.ai. Source: 8 months ago
I haven't messed with anything more fancy than Festival but I would look at coqui.ai. Source: 11 months ago
This. You can create voice models for TTS with a variety of systems - commercial and free eg https://www.resemble.ai https://coqui.ai etc - and use that with gpt text. But I don’t think you can get gpt to directly do the tts. My guess is OP accidentally made this confusing in their post title. Source: 11 months ago
>How do you detect speech starting and stopping? https://github.com/snakers4/silero-vad. - Source: Hacker News / 6 months ago
You could look into https://github.com/guillaumekln/faster-whisper especially the VAD section (Voice Activity Detector) using https://github.com/snakers4/silero-vad. Source: 10 months ago
I also had the same synchronization issue, so I wrote a WebUI/CLI that uses Silero-VAD that first splits the audio whenever there a silent portion (or every 30 seconds), and I haven't experienced it since:. Source: 11 months ago
By the way, I've updated the WebUI to now also support using Silero VAD to break up the audio into distinct sections, and run Whisper on each section and then combine them into one single transcript/SRT file. Source: over 1 year ago
And while googling this, I stumbled upon this discussion on the Whisper GitHub repository, which seems to suggest that the issue is that the current VAD (Voice Activity Detection) is quite poor, and that it can be resolved by using another VAD (like silero-vad). This might be something I want to add to my WebUI in the future. Source: over 1 year ago
Rask AI - Say goodbye to expensive translators. Our goal is to provide a dubbing and translation experience with AI that is as good as a human
Whisper Memos - Whisper Memos turns your ramblings into paragraphed articles, and emails them to you.
The Parodist App - Super-realistic celebs' voices made by AI
Lovo.ai - AI Voice Creation Platform for marketing, HR, audiobook, e-learning, movies and games.
MacWhisper - High Quality Text Transcription with OpenAI's Whisper on Mac
Fourie - A GenAI Multimodal Content Localization Platform