Back to Blog

Speech-to-Action: Voice on Your Mac Should Do More Than Type

article 5 min read

Speech-to-Action: Voice on Your Mac Should Do More Than Type

TL;DR: Speech-to-text is a solved problem. The unsolved problem is speech-to-action — taking a natural language voice command and executing it across multiple apps. Two shifts made this viable: transcription latency dropped below two seconds, and LLMs can now parse messy speech into structured intents. mrmr is a voice-first interface for Mac that does exactly this — dictate, search, and execute actions across Slack, Linear, Google Calendar, Google Meet, and Zoom from a single voice command.


You’ve tried Mac dictation. You held down the microphone key, spoke a sentence, watched it fumble half the words, and went back to typing. Maybe you tried Siri — asked it to send a message, got a “sorry, I can’t do that in this app,” and never tried again.

Most people’s relationship with voice on their computer ends there. And it’s not because voice input is a bad idea. It’s because every voice tool on macOS has the same limitation: it can only type.

That’s the ceiling. Speak, get text. Speak, get text. The entire model treats your voice as a slower, less accurate keyboard.

But your keyboard doesn’t just type. It opens apps, triggers shortcuts, switches windows, navigates menus. It’s a control layer for your entire machine. Voice should be the same — not a replacement for your keyboard, but a parallel interface that can do things across your tools.

That’s the difference between speech-to-text and speech-to-action.

The gap nobody’s closed

Speech-to-text is, for most practical purposes, a solved problem. Whisper, Deepgram, AssemblyAI — modern transcription engines are fast and accurate enough that latency and error rates are no longer the bottleneck. The problem isn’t getting words out of your mouth and onto the screen. The problem is that once the words are on the screen, you still have to do everything yourself.

Say you’re wrapping up a code review and want to flag a bug. Today, that means: open Linear, click new issue, type the title, set priority, assign it, copy the link, switch to Slack, find the channel, paste the link, type a message, send. That’s eight context switches for a task that took three seconds to say out loud.

The missing layer isn’t transcription. It’s intent parsing and execution — taking a natural language command, figuring out what you meant, identifying which apps are involved, and actually doing the thing. Not suggesting it. Not drafting it. Doing it.

What this looks like in practice

Here’s a real example. You hold fn + shift and say:

“Create a high-priority Linear ticket for the authentication timeout bug, then message the engineering channel in Slack with the ticket link.”

What happens:

Your voice gets transcribed. The system parses two distinct actions: create a Linear issue, send a Slack message. It knows “engineering channel” refers to #engineering in your workspace because it has context about your actual Slack channels and teammates. It structures both actions, shows you exactly what it’s about to do — title, priority, channel, message text — and waits.

You review. You confirm. Both actions execute. The Slack message includes the real Linear ticket link because the second action chained off the output of the first.

Total time: maybe six seconds, including the confirmation step. No window switching. No mouse. No typing.

This is what mrmr does in Action Mode. It’s not a concept or a roadmap — it’s live, working across Slack, Linear, Google Calendar, Google Meet, and Zoom today.

mrmr Action Mode

Why this didn’t work before

Siri launched in 2011. Alexa in 2014. Google Assistant in 2016. Voice assistants have been around for over a decade, and none of them became the way people get real work done. Why?

Two things were missing.

First, transcription was too slow and too wrong. If you have to wait three seconds and then correct two words, you’re slower than typing. Modern transcription engines like Whisper brought latency under two seconds and accuracy above 95% for conversational speech. That crossed the threshold where voice actually feels instant.

Second — and this is the bigger one — parsing natural language into structured actions was fragile. Old-school NLU required rigid command syntax. “Set a timer for five minutes” worked. “Remind me to check on the deployment after lunch” didn’t. The system couldn’t handle ambiguity, so you had to learn its language instead of it learning yours.

LLMs changed that. You can now take messy, natural speech — half-formed sentences, vague references, chained requests — and reliably extract structured intents from it. “Message Sarah that the PR is ready and create a ticket for the follow-up” is parseable now in a way it simply wasn’t three years ago.

These two shifts — fast transcription and flexible intent parsing — are what make speech-to-action viable. Not perfect. But viable enough that it’s faster than the alternative for a growing set of tasks.

What changes when voice becomes a control layer

If voice reliably executes actions across your tools, something fundamental shifts in how you interact with software. Your apps stop being places you go and start being services that run in the background.

Right now, using Slack means opening Slack. Using Linear means opening Linear. Your attention follows your apps. With a voice execution layer, your attention stays on your work. The apps are still there — they’re just not in your way.

This isn’t about replacing GUIs. Complex tasks still need visual interfaces. You’re not going to refactor a codebase by voice or design a slide deck by talking. But a huge portion of daily knowledge work — messages, tickets, calendar events, status updates, quick searches — is fundamentally simple. It’s just scattered across six different apps that each demand your full attention to do one small thing.

Voice-as-control-layer collapses that friction. One command, multiple apps, no switching.

The longer-term picture is even more interesting. If a voice interface has context about your tools, your teammates, and your patterns, it stops being a command executor and starts being a personal composer — anticipating what you need, chaining actions you haven’t asked for yet, and adapting to how you actually work rather than how software designers assumed you’d work.

We’re not there yet. But the foundation is being built now, and it starts with proving that voice can reliably do more than type.

Try it

mrmr is in private beta for macOS. If you want to see Action Mode firsthand, book a demo call or join the waitlist.

Join the waitlist → Book a 20-minute demo →

Get early access

mrmr is in private beta. Join the waitlist or book a demo call for fast-track access.

Stay in the loop

Get product updates and tips for voice-first workflows. No spam.



mrmr logo