@bedast so long the model is outsourced to OpenAI and the like. You can always be certain everything you ever watch on VLC will be beamed to a third party "for improvement". Auto-generated subtitles might be better than no subtitles, but not at the cost of constantly feeding third parties with your data.
And of course, if we are talking of OpenAI's models, they are known to outright invent nonsense phrases when they tried audio transcription a few months ago.
Id not trust an hallucinegic liar.
@zanagb @bedast VLC will do it in-device, not sending anything anywhere.
Whisper models are terrible at transcribing casual conversations of doctors and patients because the training data doesn't reflect that kind of speech and environments. But it excels at transcribing movies etc. because a lot of its training data are closed captions. So this would actually work reasonably well. One can put some text with the names of characters, places, etc. as context and that makes it transcribe those names very well. (source: I've been using whisper models at work, and occasionally I've been putting the mic towards the speaker with some show I'm watching to test) (also: I haven't sent any data to openai nor paid them anything)
@zanagb @bedast VLC will do it in-device, not sending anything anywhere.
Whisper models are terrible at transcribing casual conversations of doctors and patients because the training data doesn't reflect that kind of speech and environments. But it excels at transcribing movies etc. because a lot of its training data are closed captions. So this would actually work reasonably well. One can put some text with the names of characters, places, etc. as context and that makes it transcribe those names...