Methodology

Accuracy, speed, and trust signals (with receipts)

This page is the source of truth for every accuracy, speed, and language claim on ScreenApp.io. Numbers come from our internal test corpus, the Groq engineering case study, OpenAI's Whisper benchmarks, and xAI's published Grok Speech-to-Text benchmarks. Last refreshed: May 2026.

The model stack

Transcription does not bet on a single vendor. ScreenApp routes each job to the provider best suited to the audio: source platform, length, channel layout, language. As of May 2026, the primary providers are:

Fallback providers (used when a primary provider is rate-limited or unavailable): Cloudflare Workers AI, Fireworks AI, Mistral, Baseten. A transcription job never fails because one vendor had an outage.

LLM layer for summarization, chat, and AI analysis: Google Gemini end-to-end. ScreenApp is NOT powered by GPT-4, ChatGPT, or Claude. The LLM that turns a raw transcript into structured summaries, chapter markers, action items, Q&A answers, and the chat interface is Gemini from start to finish.

Why multi-provider instead of single-vendor: each provider has a sweet spot. Whisper has the most languages, Groq makes it fast. Gemini is best at short audio and downstream LLM analysis. xAI Grok STT has the lowest published error rate on phone calls. Routing each job to its best fit beats picking one provider for everything.

Customer audio is never used to train any of these models. Audio is processed and deleted per your account's retention settings. Full data handling on the Trust Center.

Speed: the Groq case study

In 2025, ScreenApp moved from a self-hosted Whisper deployment on AWS to Groq's inference infrastructure. Groq published the case study; the numbers below are from their engineering team's measurements.

MetricBefore GroqAfter GroqChange
20-minute transcription job~20 minutes~15 seconds20x faster
Per-minute transcription costbaseline1/15th15x cheaper
Free-to-paid conversionbaseline+30%uplift
Annual recurring revenue (year-over-year)baseline+405%growth attributed to the speed and cost gains

Source: ScreenApp + Groq case study (groq.com).

What this means in practice: a 60-minute meeting completes in roughly 3 minutes end-to-end (transcription, diarization, summary generation). A 2-hour video processes in about 6 minutes. These are end-to-end times that include summarization and chaptering, not just raw transcription.

Accuracy: word error rate benchmarks

Word error rate (WER) counts substitutions + deletions + insertions per 100 reference words. Lower is better. Baseline figures below come from the published benchmarks for each underlying model; the per-condition rows are from our own April 2026 retest on 18 hours of public-domain audio per language across three conditions: studio (single speaker, treated room), conference (multi-speaker, room mic), and field (handheld phone mic, ambient noise).

Published baselines

Per-language WER (April 2026 retest)

Language Locale Studio WER Conference WER Field WER iPhone mic WER † Speakers tested
English (US)en-US4.2%7.8%12.4%14.9%4
Spanish (Latin Am.)es-4195.1%9.2%14.6%17.5%3
Spanish (Spain)es-ES5.4%9.8%15.1%18.1%3
Portuguese (BR)pt-BR5.8%10.1%15.8%19.0%3
Portuguese (PT)pt-PT6.4%11.2%17.0%20.4%2
Frenchfr-FR5.9%10.4%16.2%19.4%3
Germande-DE6.1%10.8%16.5%19.8%3
Italianit-IT6.3%11.0%17.1%20.5%3
Japaneseja-JP7.8%13.5%19.8%23.8%2
Koreanko-KR7.5%13.1%19.2%23.0%2
Mandarin (Simplified)zh-CN7.9%14.0%20.4%24.5%3
Hindihi-IN9.2%15.8%23.1%27.7%3
Arabic (MSA)ar9.6%16.2%24.0%28.8%2
Russianru-RU6.8%11.5%17.4%20.9%3
Indonesianid-ID7.1%12.4%18.5%22.2%2

iPhone mic WER is a projection, not a measurement. Computed as Field WER × 1.2 to account for the noise-cancellation, beam-forming, and codec losses that iPhone built-in microphones introduce on top of a handheld phone mic. The July 2026 retest will replace this column with measured numbers from an iPhone 14, iPhone 15, and iPhone 16 across the same 18 hours per language.

Test methodology

Speaker diarization

Diarization (attaching a speaker ID to each word) runs on whichever path fits the audio. Routing is automatic; you do not pick.

The pipeline reads the audio's channel layout, source platform, and content type, then picks the better path. Word-level granularity in both cases (not paragraph-level), so a one-sentence interjection in a multi-speaker meeting gets attributed correctly.

Sources: xAI Grok STT and TTS API announcement, Google Gemini.

Supported languages

ScreenApp supports 99 languages for transcription via Whisper Large-v3. A subset of 25 of those also supports speaker diarization via xAI Grok STT (marked with †).

Full language list (Whisper Large-v3)

Afrikaans, Albanian, Amharic, Arabic †, Armenian, Assamese, Azerbaijani, Bashkir, Basque, Belarusian, Bengali, Bosnian, Breton, Bulgarian, Burmese, Cantonese, Catalan, Chinese (Mandarin) †, Croatian, Czech, Danish, Dutch †, English †, Estonian, Faroese, Finnish, French †, Galician, Georgian, German †, Greek, Gujarati, Haitian Creole, Hausa, Hawaiian, Hebrew, Hindi †, Hungarian, Icelandic, Indonesian †, Italian †, Japanese †, Javanese, Kannada, Kazakh, Khmer, Korean †, Lao, Latin, Latvian, Lingala, Lithuanian, Luxembourgish, Macedonian, Malagasy, Malay, Malayalam, Maltese, Maori, Marathi, Mongolian, Nepali, Norwegian, Nynorsk, Occitan, Pashto, Persian, Polish, Portuguese †, Punjabi, Romanian, Russian †, Sanskrit, Serbian, Shona, Sindhi, Sinhala, Slovak, Slovenian, Somali, Spanish †, Sundanese, Swahili, Swedish, Tagalog †, Tajik, Tamil, Tatar, Telugu, Thai, Tibetan, Turkish †, Turkmen, Ukrainian, Urdu, Uzbek, Vietnamese †, Welsh, Yiddish, Yoruba.

† indicates languages with word-level speaker diarization via xAI Grok STT. Other languages are transcribed (text-only) without per-speaker attribution.

Whisper's language list source: github.com/openai/whisper. Grok STT language coverage source: xAI Voice docs.

Platform availability

ScreenApp ships native apps on iOS, macOS, and Android alongside the web app. The numbers below are pulled from each platform's canonical store listing or download endpoint and verified on May 15, 2026. This section is refreshed every release cycle.

iOS app (iPhone, iPad, Apple Silicon Mac via Catalyst)

iOS App Privacy nutrition label

This is the privacy declaration ScreenApp submits to Apple, rendered exactly as it appears on the App Store. Apple's listing is authoritative; if the table below ever diverges from the live App Store page, the App Store page wins and we will update this table within 7 days. ScreenApp declares no tracking data.

Group Category Data types
Data Used to Track You None. ScreenApp does not declare any tracking data.
Data Linked to You User Content Photos or Videos, Audio Data
Identifiers User ID
Diagnostics Performance Data
Data Not Linked to You Identifiers Device ID
Contact Info Email Address, Name
Diagnostics Crash Data, Other Diagnostic Data

Canonical source: the App Privacy section on the ScreenApp App Store page. Full data handling policies on the Trust Center.

macOS (native desktop app)

Android

Rating and review counts move daily on the App Store and Google Play. The numbers on this page are point-in-time snapshots, dated above. The live store listings are always the authoritative source; if the divergence ever exceeds 0.2 stars or 10 percent of reviews, please flag it via the Trust Center contact form and we will refresh sooner.

Selected customer reviews

A small sample of named, named-role customer reviews drawn from the public reviews page. These are samples, not the full corpus. Aggregate ratings on each app remain whatever the App Store and Google Play report (4.0 stars across 94 ratings on iOS, 3.9 stars across 678 ratings on Android), not the average of the three reviews below.

Mobile and Desktop Excellence

The desktop and mobile apps are fantastic. Recording meetings while I'm mobile has never been easier, and the dictation feature is a huge time-saver. It helps me develop faster and eliminates the guesswork of recalling meeting details.

Kelvin, Software Engineer

A terrific and pleasant recording system

Our overall experience with ScreenApp has been nothing but pleasant. Their support is terrific, and ScreenApp is a great recording system.

Aaron, Verified Capterra User

Game-Changer for Client Calls

Our team was drowning in client feedback until we found ScreenApp. Now we record every presentation and client call, and the AI summaries are spot-on. My team actually looks forward to review sessions now because everything is searchable and actionable.

Millie, Director

User base

2,163,740 accounts registered as of May 2026. Number reflects unique verified-email accounts in our production database. We refresh this figure quarterly on this page.

We do not publish round-number marketing claims like "2 million users" without the verifiable underlying count, on this page or elsewhere. If you ever see an inflated or undated user-count claim on a ScreenApp page, that's a content quality issue and we'd like to know: contact us via the Trust Center.

Free access and pricing

Two ways to use ScreenApp without paying upfront:

Paid plans

Current pricing and feature breakdown on the pricing page. What we do NOT offer: a recurring monthly free tier with X minutes/month, a "no credit card" trial of the paid plan, or unlimited free recordings. If you see those claims anywhere on this site, that's a content quality issue, please flag it via the Trust Center and we'll fix the source page.

Security and compliance

SOC 2 Type 2 audited annually. 22 internal policies covering access control, data classification, secure development, and incident response. Continuous control monitoring.

Full live security posture, downloadable SOC 2 Type 2 report, and pre-filled security questionnaire at our Trust Center (trust.inc/screenapp).

Sources and external benchmarks

Errata and corrections

Numbers on ScreenApp pages should match this page. If you find a feature page that contradicts these figures, that's a content quality bug we want to fix. Report it via the Trust Center contact form and we'll update the source page within 7 days.

User
User
User
Join 2,163,740+ users

Try ScreenApp on a real recording

Paste a URL or upload an audio or video file. See the actual accuracy on your content, not benchmarks on someone else's.

Start the 7-day free trial

No charge for 7 days, cancel anytime • No credit card required