Blogged some notes on the new (still MIT licensed) Whisper Turbo model, quietly released by OpenAI yesterday
It’s both smaller and 8x faster than their previous Whisper Large https://simonwillison.net/2024/Oct/1/whisper-large-v3-turbo-model/
And you can run it on a Mac with “pip install mlx-whisper” and then:
import mlx_whisper
print(mlx_whisper.transcribe(
"path/to/audio",
path_or_hf_repo="mlx-community/whisper-turbo"
)["text"])
@simon interesting they are releasing speech to text models. I wonder will anyone be releasing improved text to speech models anytime soon?