Whisper for Mac
Running OpenAI Whisper locally on macOS — on Apple Silicon, without the cloud, without an API key, without billing. A practical guide.
Why run Whisper locally
OpenAI's Whisper is the speech recognition model that made high-quality transcription accessible to everyone. You can call it through the OpenAI API — that's the path most people take first. But the model weights are open, and on Apple Silicon you can run them locally, at native speeds, without your audio ever leaving your Mac.
There are three reasons to do this. One: privacy. Everything stays on your device — meetings, memos, sensitive conversations, the stuff you wouldn't put in a third-party tool. Two: cost. The API charges per minute; local inference on hardware you already own is free after the initial model download. Three: latency and offline. No network round-trip means faster dictation, and it works on airplanes, in air-gapped environments, anywhere you don't have a connection.
The three paths to Whisper on Mac
1. whisper.cpp (command line)
whisper.cppis Georgi Gerganov's ultra-fast C++ port of Whisper. It runs natively on Apple Silicon with Metal and Core ML acceleration, supports every Whisper model size (tiny to large-v3 and Turbo), and can transcribe audio files directly from the terminal. It's what most Mac-native Whisper tools are built on under the hood.
Good for: developers, scripted pipelines, file transcription, local servers. Not a dictation experience on its own — you still need something on top that handles hotkeys, audio capture, and pasting into the focused text field.
2. Python / OpenAI whisper (reference implementation)
The original OpenAI whisperPython package works on Mac too. It's a cleaner codebase for research and experimentation, with access to every official model variant. On Apple Silicon it's slower than whisper.cpp because it doesn't use Core ML or Metal as aggressively, but it's the reference for checking behavior and model output.
3. A native Mac app built on local Whisper
Most people don't want to run a Python script to dictate an email. They want a hotkey that pops open a mic and pastes clean text where the cursor is. That's what apps like Resonant, SuperWhisper, VoiceInk, and MacWhisper are for — they wrap local Whisper in an actual product.
Of those, Resonant is the one we build. It uses local Whisper (and Parakeet and Moonshine) models on Apple Silicon, runs system-wide through a configurable hotkey, and adds filler-word removal, sentence cleanup, and custom vocabulary on top of the raw transcription. MacWhisper focuses on file transcription; VoiceInk and SuperWhisper take different angles on the same local-Whisper foundation.
Which Whisper model should you use on Mac?
Whisper comes in several sizes — tiny, base, small, medium, large-v2, large-v3 — plus Turbo and distilled variants. Bigger models are more accurate but slower. For live dictation on Apple Silicon, Turbo and distil-large hit the best point on the speed/accuracy curve for most English use cases. For multilingual accuracy, large-v3 is still the gold standard.
Resonant handles model selection automatically and defaults to a fast, high-accuracy model tuned for dictation latency. If you're building your own pipeline, Turbo is usually the right starting point on M-series chips.
Core ML and the Neural Engine
Apple Silicon includes the Neural Engine — a dedicated co-processor for machine learning workloads. Whisper models can be converted to Core ML format to run on it, which is dramatically more power-efficient than CPU or GPU inference. whisper.cpp supports Core ML acceleration, and well-built Mac apps take advantage of it to keep dictation fast without burning battery.
The easy answer
If you want to run Whisper on Mac and you already know why — whisper.cpp is the right foundation and it's free. If you want the experience of Whisper — press a key, speak, clean text in any app — download Resonant. Same model under the hood, wrapped in a product that handles all the parts between “I have a thought” and “the text is in Slack.”
Frequently asked questions
Can I run OpenAI Whisper on Mac without the API?
Yes. Whisper's model weights are open. On Apple Silicon, whisper.cpp runs them natively with Core ML and Metal acceleration. No API key, no cloud, no per-minute billing.
Is local Whisper as accurate as the API?
Yes. Local Whisper uses the same model weights. Output is the same. The only differences are latency (local is faster — no network round-trip), cost (local is free), and privacy (local keeps audio on-device).
How do I use Whisper for live dictation on Mac?
Raw Whisper doesn't have a dictation UX — it's a batch transcription model. You need a wrapper that handles hotkeys, streaming audio, and pasting text. Resonant is a free Mac app built exactly for that, using local Whisper models.
What's the fastest Whisper model on Apple Silicon?
Whisper Turbo and distil-whisper variants are the fastest at near-equivalent English accuracy. On M-series chips with Core ML acceleration, they deliver real-time transcription comfortably.
Do I need an Intel Mac or Apple Silicon for Whisper?
Whisper runs on both, but Apple Silicon is dramatically faster because of the Neural Engine and unified memory. For practical live dictation, Apple Silicon is the recommended platform.
Try Resonant free
Private voice dictation for Mac and Windows. 100% on-device, no account required. Download and start speaking in under a minute.