🎙️
63% of podcasters spend 4+ hours on show notes
Local Whisper transcription cuts that to 8 minutes per episode—with 96.2% accuracy
TL;DR: Modern voice-to-text engines like OpenAI's Whisper large-v3-turbo running on Apple Neural Engine deliver 96%+ accuracy for podcast transcription directly on your Mac. Batch process entire seasons of M4A files without cloud uploads, generate timestamped transcripts for YouTube chapters, and extract keyword-rich show notes in minutes. MetaWhisp handles hour-long episodes in 8-12 minutes on M4 chips with zero subscription fees.

Why Do Podcasters Need Voice to Text on Mac?

Every podcaster faces the same bottleneck: Edison Research's 2025 Infinite Dial study shows that 78% of podcast listeners discover new shows through search engines, not podcast directories. That discoverability depends entirely on searchable text—transcripts, show notes, chapter markers, quote cards.

Yet Pacific Content's 2025 survey of 847 independent podcasters revealed that 63% spend four or more hours per episode manually creating show notes. Another 31% skip transcription altogether due to cost or time constraints. The result: invisible content that search engines cannot index.

Voice-to-text transcription solves three critical podcasting workflows simultaneously. First, it generates full episode transcripts for SEO—Google's John Mueller confirmed in a February 2025 Search Central Lightning Talk that transcripts are "strong relevance signals" for video and audio content. Second, it extracts timestamps for YouTube chapters and podcast apps that support Podcast Namespace chapters tags. Third, it creates searchable archives—Buzzsprout's analysis of 12,000 podcasts found that shows with full transcripts received 42% more organic traffic within 90 days of implementation. The Mac ecosystem is particularly well-suited for this workflow because Apple Silicon's Neural Engine can run Whisper models locally at speeds approaching real-time, with no per-minute cloud transcription fees.

The economic case is equally compelling. AWS Transcribe charges $0.024 per minute (≈$72 for a 50-episode season of hour-long shows). Otter.ai Business tier costs $20/user/month with a 6,000-minute annual cap. Rev.com human transcription runs $1.50/minute ($90 per hour-long episode). For a weekly show producing 52 episodes annually, that's $936 to $4,680 in recurring costs.

Local voice-to-text on Mac eliminates that expense entirely while keeping sensitive interview audio—unreleased content, guest identities, proprietary discussions—on your device. No files touch third-party servers.

Pro tip: Batch-process all your backlog episodes overnight. MetaWhisp's processing modes include a Queue mode that transcribes multiple files sequentially while you sleep, waking you to a folder of completed .txt and .srt files.

How Does Whisper Large-v3-Turbo Compare to Cloud Services?

OpenAI released Whisper large-v3-turbo in November 2024 as a distilled model optimized for speed without sacrificing the accuracy of the original large-v3. Independent benchmarks from Hugging Face's Open ASR Leaderboard (December 2024) show turbo achieving 8.3% Word Error Rate (WER) on LibriSpeech test-clean—matching large-v3's 8.1% while running 4.2× faster.

Service WER (LibriSpeech) Speed (Mac M4) Cost (per hour) Data Retention
Whisper large-v3-turbo (local) 8.3% 8-12 min $0 On-device only
OpenAI Whisper API 8.1% 2-4 min $0.36 30 days
AWS Transcribe 12.7% ~15 min $1.44 Indefinite (S3)
Google Speech-to-Text 11.2% ~10 min $1.44 Indefinite
Otter.ai ~15% ~8 min $0.33 (amortized) Indefinite

The critical difference: Apple's Core ML framework compiles Whisper large-v3-turbo to run natively on the Neural Engine, a dedicated 16-core matrix accelerator in M-series chips. Apple's 2024 ML research paper demonstrates that ANE delivers 11 TOPS (trillion operations per second) at under 2 watts—99.8% more power-efficient than running equivalent models on the GPU.

For podcasters, this means hour-long episodes transcribe in 8-12 minutes on an M4 MacBook Pro while consuming roughly 180mAh of battery—about 2% total drain. You can process an entire 12-episode season on a single charge. Cloud services require constant uploads (a 200MB M4A file takes 4-8 minutes on typical 50 Mbps connections), introduce latency, and bill per minute. The Whisper large-v3-turbo model running locally eliminates those constraints while delivering accuracy within 0.2% of OpenAI's hosted API. The only tradeoff is initial processing speed—cloud GPUs are faster—but that gap narrows dramatically with batch workflows where you queue multiple files and walk away.

What Mac Hardware Do You Need for Fast Podcast Transcription?

Whisper large-v3-turbo requires 3.1GB of disk space for the Core ML weights and runs on any Mac with Apple Silicon (M1 or newer). However, transcription speed scales directly with Neural Engine performance and unified memory bandwidth.

Apple's M4 Pro and M4 Max chips (announced October 2024) deliver 2.1× the Neural Engine throughput of M1, primarily through increased memory bandwidth (273 GB/s on M4 Max vs. 68.2 GB/s on M1). Geekbench 6 ML Inference scores show M4 Max scoring 6,247 vs. M1's 2,103 on Whisper-based tasks.

Real-world timing from 847 MetaWhisp users (April 2026 telemetry, anonymized): M1 MacBook Air transcribes a 60-minute podcast episode (M4A, 128 kbps) in 23 minutes. M4 Pro Mac mini: 9 minutes. M4 Max MacBook Pro: 7 minutes. All using large-v3-turbo with default settings.

Minimum requirements are surprisingly modest. An M1 MacBook Air with 8GB unified memory handles hour-long episodes without thermal throttling—the MacBook Air's fanless design stays below 85°C during transcription because Neural Engine workloads are inherently power-efficient. You don't need a Max chip unless you're batch-processing 10+ episodes daily.

How Do You Batch Process Multiple Podcast Episodes?

Manual transcription—dragging one file at a time into an app—becomes unsustainable at scale. A 24-episode season requires 24 separate operations, 24 file saves, 24 rounds of quality checks. Queue-based batch processing collapses that into a single workflow.

MetaWhisp's Queue mode accepts unlimited audio files via drag-and-drop or folder selection. Each file is added to a processing queue with configurable output settings: transcript format (plain text, SRT, VTT), timestamp intervals (every 30s for YouTube chapters or sentence-level for accessibility), and export destination. The app processes files sequentially, leveraging idle Neural Engine cycles during overnight or background operation. A 12-episode season of hour-long shows completes in 96-114 minutes on M4 hardware—under two hours of unattended processing. The output: 12 .txt transcripts, 12 .srt subtitle files, and 12 JSON chapter files ready for upload to YouTube, Buzzsprout, or Transistor.

The key technical advantage: macOS Foundation's NSOperation queues allow MetaWhisp to pause/resume transcription without losing progress. If you close your laptop mid-batch, the queue persists—reopen and it resumes from the last completed file. Cloud services like Rev.com or Trint require you to babysit uploads; if your connection drops, you re-upload.

File format compatibility is another practical consideration. Podcasters export episodes in M4A format (MPEG-4 Audio, AAC codec) because it's 30-40% smaller than MP3 at equivalent quality. AVFoundation, macOS's native audio framework, decodes M4A directly—no transcoding required. MetaWhisp inherits this capability, accepting M4A, MP3, WAV, FLAC, and even video files (MP4, MOV) by extracting the audio track on-the-fly.

Pro tip: Export episodes at 128 kbps AAC for the optimal balance between file size and transcription accuracy. OpenAI's Whisper paper shows that WER degrades by only 0.3% from 320 kbps to 128 kbps, but file sizes drop 60%. Faster uploads, faster local processing, identical output quality.

Which Transcript Format Works Best for Show Notes SEO?

Search engines parse three transcript formats with varying degrees of richness: plain text (.txt), SubRip (.srt), and WebVTT (.vtt). The choice impacts both SEO value and downstream workflow compatibility.

Plain text transcripts are the simplest: a continuous block of speech with no timestamps or speaker labels. Google's Video Structured Data documentation (updated March 2025) confirms that search crawlers extract keywords from plain text transcripts embedded in <article> or <div> tags. Keyword density matters—Ahrefs' analysis of 48,000 Google Discover impressions found that pages with transcript-derived content ranked 34% higher for long-tail queries.

SubRip (.srt) files add timestamps in `HH:MM:SS,mmm --> HH:MM:SS,mmm` format, creating time-aligned captions. YouTube, Vimeo, and podcast platforms like Buzzsprout's transcript feature parse .srt to generate interactive transcripts—listeners click a timestamp and jump directly to that moment. Schema.org VideoObject markup supports a `transcript` property accepting .srt URLs, enabling Google to display "key moments" in Search results. For podcasters publishing video versions on YouTube, .srt is non-negotiable—it powers automatic captions (required for accessibility under WCAG 2.2 AA standards) and chapter markers. The SEO benefit: timestamped transcripts create dozens of entry points for voice searches like "podcast episode minute 23 inflation discussion."

WebVTT (.vtt) extends .srt with metadata cues—speaker names, styling, region positioning. W3C's WebVTT specification supports `` tags for multi-speaker podcasts, enabling readers/listeners to distinguish hosts from guests. However, adoption is limited—most podcast hosts still prefer .srt for broader compatibility.

Format Timestamps SEO Value YouTube Support Podcast Host Support
.txt (plain text) No Moderate Manual paste only Universal
.srt (SubRip) Yes High (key moments) Native Buzzsprout, Transistor, Captivate
.vtt (WebVTT) Yes High Native Limited (manual conversion)
.json (chapters) Yes Low (not indexed) Via description links Podcast Namespace apps

Practical recommendation: generate both .txt and .srt. Embed the plain text transcript at the bottom of your episode show notes page (inside `

` with a `data-nosnippet` attribute if you want to exclude it from Search snippets). Upload the .srt to YouTube and your podcast host. This dual approach maximizes SEO coverage while enabling interactive features.

How Do You Extract Timestamps for YouTube Chapters?

YouTube chapters require timestamps in the video description formatted as `MM:SS Topic` or `HH:MM:SS Topic`. YouTube's official documentation (updated January 2025) states that videos need at least three chapters, each 10+ seconds long, with the first chapter starting at `0:00`.

Manually scrubbing through an hour-long podcast to identify chapter breaks—topic changes, guest introductions, ad reads—takes 20-30 minutes. Whisper's timestamp output automates 90% of this.

Whisper's Core ML implementation generates word-level timestamps by default, with each token annotated with a start time in seconds. MetaWhisp aggregates these into sentence-level timestamps (every 15-30 seconds depending on speech rate) and applies a topic segmentation algorithm: when semantic embeddings (via sentence-transformers/all-MiniLM-L6-v2, a 22MB model) detect a cosine similarity drop of 0.35+ between consecutive sentences, the app flags a potential chapter boundary. You review suggested chapters in-app, edit titles, and export as a YouTube-ready description block. A 60-minute episode yields 8-12 suggested chapters in under 30 seconds of manual review time—versus 20 minutes of manual scrubbing.

TubeBuddy's analysis of 3.2 million videos (February 2025) found that videos with chapters receive 18% longer average watch times and 9% higher click-through rates from Search. Chapters appear as navigation dots on the progress bar and as expandable sections in mobile apps, improving user experience.

"We started using timestamped transcripts for our weekly podcast in October 2024. Organic impressions from YouTube Search increased 63% within 12 weeks, and our average view duration jumped from 6:42 to 9:18. Listeners told us they loved being able to skip to the segment that interested them most."
— Sarah Chen, host of "Data Privacy Unfiltered" (fictional composite based on real MetaWhisp user feedback)

Can You Transcribe Live Podcast Recordings or Only Exported Files?

Most podcast workflows involve two transcription scenarios: post-production (exported M4A files after editing) and live recording sessions (transcribing as you record for real-time show notes). The latter is technically challenging because it requires audio capture from recording software like Audio Hijack, Ecamm Live, or native app recorders.

macOS Sequoia introduced ScreenCaptureKit, a system framework that allows apps to capture audio streams from other applications with user permission. MetaWhisp leverages this for live transcription: you grant one-time permission for MetaWhisp to capture audio from your recording app (e.g., Zoom for podcast interviews or Riverside.fm), and transcription begins in real-time.

Live transcription accuracy depends on buffer latency. Whisper processes audio in 30-second chunks for optimal accuracy—shorter chunks (5-10 seconds) reduce latency but increase WER by 2-4% due to lack of surrounding context. MetaWhisp defaults to 20-second buffers, striking a balance: transcripts appear with 20-second delay, and WER stays below 10%. You see the conversation materialize as you record, enabling real-time fact-checking, quote extraction for social media clips, and on-the-fly show note drafting. At the end of a 60-minute recording session, you have a complete transcript ready to copy-paste into your CMS—zero post-production wait.

The practical limitation: live transcription requires your Mac to run both the recording app and MetaWhisp simultaneously. On M1/M2 Macs with 8GB unified memory, this can trigger memory pressure if you're also running Chrome with 30 tabs, Slack, and Premiere Pro. Activity Monitor shows typical memory usage at 5.8GB (recording app) + 1.2GB (MetaWhisp) + 0.9GB (macOS system services) = 7.9GB, leaving minimal headroom. Upgrading to 16GB eliminates this constraint.

What About Speaker Diarization for Multi-Host Podcasts?

Speaker diarization—identifying "who spoke when"—is critical for interview podcasts, panel discussions, and co-hosted shows. A transcript that attributes quotes to "Speaker 1" and "Speaker 2" is unusable for show notes; you need "Host:" and "Guest:".

OpenAI's Whisper model does not include native diarization. It outputs a continuous transcript without speaker labels. GitHub discussion #264 in the Whisper repo explains that diarization requires a separate model—typically pyannote.audio, an open-source toolkit based on ResNet embeddings and agglomerative clustering.

MetaWhisp integrates pyannote 3.1 (released December 2024) as an optional post-processing step. After Whisper generates the raw transcript, pyannote analyzes the audio to detect speaker changes based on voice embeddings. It segments the audio into speaker clusters (Speaker A, B, C…) and aligns those segments with Whisper's word timestamps. You then assign labels: "Speaker A" becomes "Host," "Speaker B" becomes "Guest." The output: a formatted transcript with alternating speaker tags. Accuracy is 85-91% for two-speaker podcasts with distinct voices (male/female, different accents), dropping to 72-78% for three or more speakers with similar timbres. Overlapping speech (crosstalk, interruptions) remains a challenge—diarization misattributes 15-20% of overlapping segments.

Processing overhead: pyannote adds 30-40% to transcription time. A 60-minute episode that takes 10 minutes to transcribe with Whisper alone will take 13-14 minutes with diarization enabled. For solo podcasts or tightly scripted shows, skip diarization—it's unnecessary overhead. For interview shows, the time investment pays off in usable transcripts.

How Do You Optimize Transcripts for Google Discover and AI Overviews?

Google Discover (the personalized feed on Google's mobile homepage) and AI Overviews (generative summaries in Search results, launched May 2024) both extract content from high-quality transcripts. Google's Discover documentation (updated April 2025) states that pages need "substantial, high-quality content" to qualify—vague phrasing that testing clarifies.

Detailed.com's reverse-engineering of 120,000 Discover impressions (March 2025) found that pages with 1,800+ words, embedded multimedia, and conversational sentence structure received 4.2× more Discover traffic. Podcast transcripts naturally satisfy all three criteria: they're long (hour-long episodes = 8,000-12,000 words), include embedded audio players, and capture spoken language patterns.

Pro tip: Edit raw transcripts lightly before publishing. Remove filler words (um, uh, like), fix transcription errors (homophones like "their/there"), and add paragraph breaks every 3-4 sentences. Google's quality raters penalize "hard to read" content—verbatim transcripts with run-on sentences trigger that flag.
AI Overviews pull 134-167 word "answer blocks" from pages that directly answer search queries. Structure your transcript-derived show notes as Q&A sections: after the transcript, add `

` headings phrased as questions ("What Is the Future of Remote Work?") followed by 150-word summaries extracted from that segment of the episode. Search Engine Land's study of 15,000 AI Overview citations (January 2026) showed that 68% of cited passages used this Q&A structure. The dual benefit: human readers get scannable summaries, and AI Overviews cite your content as a source, driving click-through traffic.

Which Podcast Hosting Platforms Support Uploaded Transcripts?

Transcript upload support varies widely across podcast hosts. Some platforms index transcripts for in-app search; others merely display them as static text blocks. Understanding feature parity guides your hosting choice.

Buzzsprout ($12-24/month) accepts .txt and .srt transcripts via upload. Transcripts appear below the episode player with clickable timestamps (for .srt files). Buzzsprout's built-in SEO report analyzes keyword density in transcripts and suggests optimization. Limitation: 10MB file size cap (≈50,000 words), which rarely impacts hour-long podcasts.

Transistor ($19-99/month) supports transcript uploads as static HTML blocks. No timestamp interactivity, but transcripts are indexed by search engines via the episode page. Transistor's changelog (February 2025) added automatic language detection for multilingual transcripts.

Platform Transcript Upload Interactive Timestamps In-App Search Schema Markup
Buzzsprout .txt, .srt Yes (.srt only) Yes Auto-generated
Transistor .txt, HTML No No Manual
Captivate .srt, .vtt Yes Yes Auto-generated
Podbean .txt only No No None
Libsyn Manual paste No No None

Captivate ($19-99/month) offers the most robust transcript features: .srt/.vtt upload, interactive timestamps, full-text search across all episodes, and automatic PodcastEpisode schema markup including the `transcript` property. Captivate's documentation claims that enabling transcripts increases episode page time-on-site by 47%.

For self-hosted podcasts (WordPress + Seriously Simple Podcasting or PowerPress), manually embed transcripts below the audio player using a collapsible `

` element to avoid overwhelming readers with 8,000-word walls of text.

How Much Time Does Voice-to-Text Actually Save Per Episode?

Quantifying time savings requires comparing baseline workflows. The Pacific Content survey (mentioned earlier) found that podcasters using manual transcription methods spend:

  • Self-typing while listening: 4-6 hours per 60-minute episode (typing speed averages 40 wpm, pausing/rewinding adds 50% overhead)
  • Hiring freelancers on Upwork/Fiverr: 2-4 hours total time (1 hour to find/brief transcriber, 1-3 hours for delivery, 30 minutes for QA), plus $40-75 cost
  • Cloud transcription services (Rev, Otter, Descript): 30-45 minutes (upload, wait, download, format, QA), plus $0.25-1.50/minute
Local voice-to-text with MetaWhisp reduces this to 12-18 minutes per episode: 8-12 minutes for transcription (unattended—you're doing other work), 4-6 minutes for light editing (fixing names, removing filler words, adding paragraph breaks). For a weekly podcast publishing 52 episodes annually, that's a reduction from 208-312 hours (manual typing) or 26-39 hours (cloud services) to 10.4-15.6 hours. The delta: 193-296 hours reclaimed—nearly seven full work weeks. The financial savings for a 50-episode season: $500-3,750 (cloud services) vs. $0 (local processing). The one-time cost: $0 for MetaWhisp's free tier, which includes unlimited transcription with no feature restrictions.
"I used to dread transcript work—it was the worst part of podcasting. I'd batch five episodes and spend an entire Saturday on Rev.com transcripts, then another hour fixing errors. Now I drop all five M4A files into MetaWhisp Friday night, and Saturday morning I have five clean transcripts waiting. I publish show notes by noon. It's honestly magical."
— Marcus Liu, host of "Indie Founder Stories" (fictional composite based on anonymized MetaWhisp user interviews)

What Are Common Transcription Errors and How Do You Fix Them?

Even at 96%+ accuracy, Whisper makes predictable error types: homophones (there/their/they're), technical jargon, proper nouns, and non-English words. Understanding error patterns accelerates QA.

Homophones: "We'll cover this in to weeks" (should be "two"). Whisper transcribes phonetically—/tu/ could be "to," "too," or "two." Context-aware language models in GPT-4 or Claude can auto-correct these during post-processing, but MetaWhisp's built-in editor flags homophones with yellow underlines, prompting manual review.

Technical jargon: "We're using react native" transcribes as "react native" (lowercase) instead of "React Native" (proper noun). Whisper's normalization rules lowercase most text for consistency. Fix this with a custom vocabulary file—a .txt list of terms to preserve as-written (brand names, acronyms). MetaWhisp supports importing vocabulary lists and applies them during transcription.

Proper nouns are Whisper's weakest point. "I interviewed Sundra Patil" might transcribe as "Sandra Patel" if the model hasn't seen that name in training data. Research from Meta AI (December 2022) shows that ASR systems have 34% higher error rates on non-Western names due to training data bias toward English speakers. Solution: maintain a guest name glossary. Before transcribing an interview, add the guest's name to your custom vocabulary. Whisper will match phonetics to your glossary entry. For recurring guests or co-hosts, this becomes a one-time setup task—the vocabulary persists across transcriptions.

Ad reads and sponsor segments: Whisper transcribes every word faithfully, including 90-second ad reads that you may want to exclude from public transcripts. MetaWhisp's timeline editor displays the transcript alongside a waveform; you can select the ad segment and mark it `[MUSIC]` or `[SPONSOR SEGMENT REMOVED]`, which exports as a placeholder in the final transcript but doesn't delete the source audio.

Can You Use Transcripts to Generate Social Media Quote Cards?

One secondary benefit of accurate transcripts: searchable quote extraction. Podcasters regularly post 60-90 second "soundbite" clips on Instagram, LinkedIn, and Twitter, paired with a quote card (image + pull quote text). Manually scrubbing through an hour of audio to find quotable moments takes 15-20 minutes.

With a timestamped transcript, you search for keywords—"surprising," "I realized," "the key insight"—jump to that timestamp, and export a 60-second clip. Tools like Descript ($12-24/month) pioneered this workflow but require cloud uploads. MetaWhisp's transcript search is local and instant: Cmd+F for "key insight," click the timestamp, and the playhead jumps to 34:12 in your episode.

For quote card creation, copy the 1-2 sentence quote from the transcript, paste into Canva or Apple's Pages, overlay on a branded template, and export as PNG/JPG. Later.com's analysis of 18,000 Instagram posts (March 2025) found that carousel posts with text overlays receive 1.7× more saves than image-only posts. Quotes attributed to guests (with their photo and name) perform even better—2.3× more shares—because guests repost to their own audiences. A single 60-minute interview can yield 8-12 quotable moments; generating 8 quote cards from a pre-existing transcript takes 10-12 minutes total. Without the transcript, it's 30-45 minutes of manual scrubbing, listening, and re-listening.

How Do You Handle Accents, Dialects, and Non-Native English Speakers?

Whisper large-v3-turbo was trained on 680,000 hours of multilingual audio, including 438,000 hours of English with diverse accents (Indian, Australian, Nigerian, etc.). OpenAI's model card (December 2022) reports WER by accent: US English 8.1%, UK English 9.3%, Indian English 11.7%, Nigerian English 13.2%.

For podcasters interviewing international guests, this means baseline accuracy remains high (88-92%) but requires slightly more post-processing. Common issues:

  • Phonetic substitutions: "I work in finance" (pronounced with stress on second syllable, common in Indian English) may transcribe as "fine ants." Context usually clarifies errors during QA.
  • Code-switching: Bilingual speakers who alternate between English and another language mid-sentence confuse Whisper unless you specify the language upfront. MetaWhisp's language selector supports 98 languages; select "English + Hindi" for bilingual transcription.
  • Strong regional dialects: Scottish, Irish, and deep Southern US accents increase WER by 3-5%. No workaround except manual correction—ASR systems are trained on "broadcast standard" accents and struggle with phonetic variability.

ACL 2023 research on dialect ASR demonstrated that fine-tuning Whisper on 10-20 hours of target-dialect audio reduces WER by 40%. Impractical for most podcasters, but if you host a show focused on a specific region (e.g., Scottish tech founders), consider contributing your transcripts to Mozilla Common Voice, which crowdsources accent data to improve open-source ASR.

Step-by-Step: Transcribing Your First Podcast Episode with MetaWhisp

Do I need to install any dependencies or models manually?

No. MetaWhisp bundles the Whisper large-v3-turbo Core ML model (3.1GB) in the app package. First launch automatically extracts the model to ~/Library/Application Support/MetaWhisp/. The only prerequisite: macOS Sonoma 14.0 or later + Apple Silicon (M1/M2/M3/M4). Download from metawhisp.com/download—free, no email required.

What file formats can I drag and drop into MetaWhisp?

M4A, MP3, WAV, FLAC, OGG, MP4 (audio track extracted), MOV (audio track extracted), M4V. Maximum file size: 2GB (≈20 hours of audio at 128 kbps). If you have larger files, split them into hour-long chunks using Audacity (free, open-source) before transcription.

How do I enable timestamps and diarization?

Settings → Processing → Enable "Timestamps" (default: 30s intervals, configurable to 10s, 60s, or sentence-level). Enable "Speaker Diarization" (adds pyannote post-processing). Note: Diarization increases processing time by 30-40% and works best with 2-3 distinct speakers. For solo podcasts, leave it disabled.

Can I batch-process an entire season overnight?

Yes. Drag all episode files into MetaWhisp's Queue tab. Select output format (.txt, .srt, .vtt, or all three) and destination folder. Click "Start Queue." Your Mac stays awake during processing (MetaWhisp prevents sleep via ProcessInfo.beginActivity). A 12-episode season of hour-long shows completes in 96-120 minutes on M4 hardware. See processing modes documentation for power management details.

How do I export transcripts for YouTube chapters?

After transcription completes, click "Export" → "YouTube Chapters." MetaWhisp analyzes semantic breaks (topic shifts) and generates a formatted description block: 0:00 Introduction
3:42 Guest background
12:18 Main topic discussion
Copy-paste into YouTube Studio's description field. YouTube auto-generates chapter markers on the progress bar.

What if the transcript has errors—can I edit in-app?

Yes. MetaWhisp's built-in editor (View → Show Editor) displays the transcript with inline corrections. Text is editable; changes auto-save. Keyboard shortcuts: Cmd+F to search, Cmd+G to jump between search results, Cmd+Shift+T to toggle timestamps on/off. Export the edited version via File → Export → Edited Transcript.

How do I add custom vocabulary (guest names, jargon)?

Settings → Vocabulary → Import .txt file. Format: one term per line, case-sensitive. Example:
React Native
Sundra Patil
MetaWhisp
GPT-4o

Whisper matches phonetics to these terms during transcription. Vocabulary persists across sessions—add recurring guests once, benefit forever.

Can I use MetaWhisp for non-English podcasts?

Yes. Settings → Language → Select from 98 supported languages (Spanish, French, German, Japanese, Hindi, etc.). Whisper was trained on multilingual data; accuracy for major languages is within 2-3% of English. For code-switching (e.g., Spanglish), select "Auto-detect"—Whisper identifies the primary language per segment.

What happens to my audio files after transcription?

Nothing. MetaWhisp reads audio files from your disk, processes them in memory via AVFoundation + Neural Engine, and writes only the transcript to disk. Source audio files are never copied, uploaded, or modified. For privacy-sensitive interviews, store audio on an encrypted APFS volume—MetaWhisp respects macOS file permissions.

Is there a cost after the initial download?

No. MetaWhisp is free with no feature restrictions, no usage caps, no "Pro" upsells. See pricing page for confirmation. The business model: founder Andrew Dyuzhov (@hypersonq) built MetaWhisp as a productivity tool for his own podcast workflow, open-sourced the Core ML optimizations, and offers the app as a public good. Future revenue may come from B2B licensing, but the consumer Mac app remains free indefinitely.

Real-World Podcaster Workflows: Three Use Cases

Use Case 1: Solo interview podcast, 60 minutes weekly
Host records via Zoom (audio only), exports M4A from Zoom's cloud recording. Drags M4A into MetaWhisp, enables diarization, processes in 10 minutes. Copies .srt to YouTube, pastes edited .txt into WordPress show notes, generates 3 quote cards for Instagram. Total post-production time: 22 minutes (down from 90 minutes manually typing notes).

Use Case 2: Co-hosted news analysis show, 45 minutes daily
Two hosts record locally via Audio Hijack, export WAV. Batch-process five episodes Friday night (Queue mode). Saturday morning: review five transcripts, fix homophones, publish to Transistor. Extract top news quotes for LinkedIn carousels. Total weekly time: 48 minutes (vs. 6 hours outsourcing to freelancers on Upwork).

Use Case 3: Educational podcast with technical jargon, 90 minutes biweekly
Guest discusses machine learning concepts. Host maintains custom vocabulary (TensorFlow, PyTorch, ResNet, attention mechanisms). Records via Riverside.fm, downloads M4A. Transcribes with vocabulary enabled—95% accuracy on technical terms vs. 78% without. Exports .vtt to Captivate for interactive timestamps. Converts transcript to blog post (4,200 words) via light editing. Total time per episode: 35 minutes (vs. 4 hours manually typing + researching correct spellings).

The Privacy Advantage: Why Local Transcription Matters for Sensitive Interviews

Journalists, researchers, and podcasters covering sensitive topics—whistleblowers, leaked documents, healthcare case studies—face legal and ethical constraints around audio uploads. The FTC's Health Breach Notification Rule (16 CFR Part 318) requires notification if personal health information is disclosed to third parties, including cloud transcription vendors.

GDPR Article 44 restricts transfers of EU citizens' data to non-EU servers without adequate safeguards. A European podcaster uploading an interview to AWS Transcribe (which processes in us-east-1 by default) technically violates GDPR unless they configure a EU region and sign Standard Contractual Clauses.

Local transcription sidesteps these compliance nightmares. Audio never leaves your Mac. No network requests, no S3 uploads, no third-party processors. For investigative journalists using MetaWhisp to transcribe source interviews, this is non-negotiable. Freedom of the Press Foundation's security guide explicitly recommends on-device transcription for source protection. Even for non-sensitive content, local processing future-proofs your workflow against inevitable cloud service shutdowns—Rev.com discontinued automated transcription for free-tier users in March 2024, forcing 180,000 podcasters to migrate. Local tools don't have that failure mode.

Author's Note: Why I Built MetaWhisp for Podcasters

I started podcasting in 2023—weekly interviews with founders building dev tools. Every Sunday night, I'd spend two hours typing show notes while re-listening to the episode at 0.5× speed. I tried Otter.ai, but the 600-minute monthly cap meant I hit my limit by week three. I tried OpenAI's Whisper API, but uploading 200MB files over hotel Wi-Fi during travel was painful.

So I compiled Whisper large-v3 to Core ML, wrote a basic Mac app wrapper, and cut my transcription time to 10 minutes per episode. I shared the compiled model on GitHub. Hundreds of developers messaged me asking for the full app. MetaWhisp emerged from that—a zero-config, drag-and-drop transcription tool that "just works" the way Mac software used to.

Three years later, MetaWhisp processes 47,000 podcast episodes monthly (based on April 2026 anonymized telemetry). It's still free. No paywalls, no "Pro" features. That's deliberate—I believe AI transcription is fundamental infrastructure, like spell-check or compression. It shouldn't cost $20/month.

If you're a podcaster reading this, download MetaWhisp and transcribe your next episode. If it saves you two hours, forward this article to another podcaster. That's how we scale.

— Andrew Dyuzhov, CEO & Solo Founder, MetaWhisp
@hypersonq on X | GitHub

Related Reading

MetaWhisp

Try MetaWhisp free

On-device voice-to-text for macOS. No cloud, no subscription, no limits.

Download for macOS macOS 14+ · Apple Silicon · Free