Feb 16, 2026

Voice Thinking: Real Life Scenarios

I’ve made 781 recordings. Here are the scenarios I face daily and how they actually work.

Scenario 1: I Copied 3 Things While Talking — AI Got Them in Order

I’m describing a feature to AI. While talking I see a screenshot in Chrome — I copy it. I keep talking. I see a code block in Cursor — I copy it. I keep talking. I find a link — I copy it.

I finish talking. I look at the transcript:

"let's take this feature from the panel"
[screenshot: Chrome panel UI]
"and use this file structure"
[code: Sources/Module/Features.swift]
"here's the reference"
[link: github.com/yemreak/optionOS/...]

Text + screenshot + code + link — in exactly the order I said them.

Inline copies within speech — text, image, code all in order
3 copies while talking — all in the order I said them

Video

I give this to AI. “That” and “this” don’t get lost. It sees what I was showing when I said what.

Try doing this after the fact. Open 3 apps, paste one by one, remember the order, rebuild the context. Impossible.

Scenario 2: I Said “Xcode” — My Xcode Notes Appeared

I’m talking about a build issue. I say “…so the Xcode build system needs to…”

A popup appears. My Xcode skill document. Automatically.

Xcode note suggested while speaking
I said "Xcode" — my Xcode document surfaced automatically

I didn’t search. I didn’t open the Skills panel. The keyword “Xcode” in my speech matched the keyword in my document’s frontmatter.

I click it. Preview shows the content. I copy the relevant section. Paste into AI.

Keyword matching — automatic suggestions from your notes
Keywords in my speech → matched against my 127 documents → suggested in real-time

This is the bridge between Dictation and Skills. You don’t need to remember 127 document names. Just talk. The right ones come to you.

Scenario 3: 10 Minutes of Talking, 0 Seconds of Waiting

I talk for 10 minutes about an architectural decision. Normally: finish talking, wait 10 minutes for transcription.

Not here. I talk for 5 seconds, pause to think for 2 seconds — that pause triggers transcription. By the time I continue talking, the first chunk is already text.

[talking]     "the module should handle its own state..."
[2 sec pause] → transcribes everything up to this point
[talking]     "...and the UI layer only reads from it..."
[2 sec pause] → transcribes this chunk too
[talking]     "...so the boundary is clear"
[done]        → last chunk transcribes
              → TOTAL WAIT: ~2 seconds (last chunk only)

Your thinking pause = processing moment. While you think, it works. By the time you finish, text is already ready.

Scenario 4: New Idea While Transcription Was Running

I just finished a 5-minute recording. Transcription is at 84%. A new idea explodes in my head. It’ll vanish in 30 seconds.

I hit ⌥A. Start talking. The previous transcription continues. The new recording starts on top.

Parallel recording — first at 84%, second just started
First recording: 84% transcribed. Second recording: just started. Both running.

When both finish, they merge. Order isn’t broken. First recording first, second recording second. Add as many as you want — they’re all part of the same flow.

Scenario 5: “I Said This While Cursor Was Open” — Found It

779 recordings. I need the one where I was talking about the search engine while Cursor was open. Last week sometime.

I open the search panel. Type “search engine”. Filter by app: Cursor. Filter by time: this week.

Dictation History — 779 recordings, filter by app, search within text
779 recordings. Found it: "search engine" + Cursor + this week

Human memory works with context: “where was I when I said that?” Date, duration, app, text — all filterable. The context you remember is the context you search by.

Scenario 6: Live Subtitles — I See What I’m Saying

I’m talking and wonder: “did I already say this?” Or I want to check if the transcription is converting correctly.

I hover over the panel. Live transcription appears on screen. Words flow as I speak.

Live transcription while speaking — subtitle on screen
Like YouTube subtitles — but for your own speech, in real-time

No need to stop, go back, check. I see it while I think. If something came out wrong, I correct it right there in the next sentence.


⌥A              → start/stop recording
⌥Escape         → cancel (deliberate — Escape does nothing)

What happens automatically:

2 sec pause     → transcribes current chunk (no waiting)
⌘C while rec    → copy added inline (in speech order)
keyword match   → related skill suggested (from frontmatter)

This Post Was Made by Voice

This blog was made by voice

Related posts

Clipboard: Real Life → 5 Agents, 1 Voice: Complete Workflow Guide → Skills Search: Real Life → Copy from iPhone + iPad → Agent Fixes It →