Swap Whisper for Avalon in two lines

The most accurate speech API for developers

Avalon is the drop-in API tuned for how developers talk. Change two lines to upgrade from Whisper and get 97.3% accuracy on AISpeak—no hallucinated model names, no misheard commands, no rewrite of your stack.

  • OpenAI-compatible endpoints for streaming and batch transcription
  • Free until October 15th, 2025 — then $0.39 per hour of audio

AISpeak

97.3% accuracy

Avalon nails AI jargon, CLI commands, and model names. Whisper Large v3 hits 65.1%.

OpenASR

7 / 8 splits won

Leads Whisper Large v3 and ElevenLabs Scribe across public benchmarks developers trust.

Drop-in

2-line swap

Swap your base URL and model name. Keep your auth headers, streaming logic, and tooling.

Why Avalon

People don't speak like an audiobook when they're prompting Claude Code or narrating a deploy. Most training data does. Avalon was trained on real developer workflows—pairing transcripts with CLI sessions and IDE captures—so it keeps the exact command, casing, and model number you said.

Benchmarking alone misses the point. We built Avalon after watching other models fumble obvious technical nouns—turning "zshrc" into "C sharp C" or hallucinating entirely new model names. Avalon stays literal, so your copilots, support analytics, and live demos sound like your users actually talk.

Stop hallucinations on AI terms

AISpeak is our benchmark of real-world clips where people say things like "Claude Code," "GPT-4o-mini," and "o3." Avalon transcribes the key term correctly 97.3% of the time. Whisper Large v3 misses it in more than one out of three attempts.

Accuracy on Coding and AI Terms
Higher is better
AISpeak-10
97.4%AvalonTop score
  • ElevenLabs Scribe78.8%
  • Whisper Large v365.1%
  • Voxtral Mini 3B59.5%
  • NVIDIA Canary 1B51.5%
AISpeak-50
97.5%AvalonTop score
  • ElevenLabs Scribe86.7%
  • Whisper Large v382.4%
  • Voxtral Mini 3B79.4%
  • NVIDIA Canary 1B71.8%
AISpeak-500
95.8%AvalonTop score
  • ElevenLabs Scribe87.5%
  • Whisper Large v384.9%
  • Voxtral Mini 3B82.9%
  • NVIDIA Canary 1B74.1%

Whisper hallucinates. Avalon doesn't.

Avalon
I've tried running this with GPT-4o, GPT-4.1, and o3.
Whisper
I've tried running this with GPT-4.0, GPT-4.1, and GPT-03.
Avalon
Claude 4 announcement.
Whisper
Claude for announcement.
Parakeet
cloud floor announcement.
Avalon
I just tried running this with GPT-4o and it didn't go very well.
Whisper
I just tried running this with GPT-400, and it didn't go very well.
Avalon
Ultrathink.
Whisper
Ultrathing.
Clips pulled from AISpeak evaluations

Proven on industry benchmarks

Avalon leads Whisper Large v3, ElevenLabs Scribe, and AssemblyAI across OpenASR. The same accuracy shows up in production: teams building copilots, onboarding agents, and support analytics already rely on Avalon to keep transcripts literal.

WER by dataset — Avalon vs. Canary 1B, Scribe, Voxtral, Whisper L3, AssemblyAI

Lower is better. Avalon wins the majority of public datasets developers reference when validating speech models.

  • Beats Whisper Large v3 on 7 of 8 OpenASR splits developers cite most often.
  • Tops ElevenLabs Scribe and AssemblyAI on the same public benchmarks.

Change two lines, keep your workflow

Avalon mirrors OpenAI's Whisper API — same request format, same response shape. Point your existing integration at the Avalon endpoint and keep shipping.

from openai import OpenAI

client = OpenAI(
    api_key="your-avalon-api-key",
    base_url="https://api.aqua.sh/v1"
)

audio_file = open("speech.mp3", "rb")
transcript = client.audio.transcriptions.create(
    model="avalon-1",
    file=audio_file
)

print(transcript.text)
Try in Playground →SDKs for Node, Python, and cURL with streaming built in.

Simple pricing

Avalon API

Everything unlocked. No minimums.

$0.39 / hour of audio
  • Billed per second, streaming or batch
  • No overages, no enterprise gatekeeping
  • Includes speaker labels and timestamps
🎉 Free until October 15th, 2025

Start building with Avalon

Spin up your Avalon API key and launch features where transcripts stay literal—in demos, copilots, and support workflows.