Try the latest GLM-4.6 with extended context, superior coding, and refined intelligence. Now available on-demand

OpenAi Logo MArk

Whisper V3 Turbo

Whisper large-v3-turbo is a finetuned version of a pruned Whisper large-v3. In other words, it's the exact same model, except that the number of decoding layers have reduced from 32 to 4. As a result, the model is way faster, at the expense of a minor quality degradation.

Fireworks Features

Serverless

Immediately run model on pre-configured GPUs and pay-per-token

Learn More

On-demand Deployment

On-demand deployments give you dedicated GPUs for Whisper V3 Turbo using Fireworks' reliable, high-performance system with no rate limits.

Learn More

Whisper V3 Turbo FAQs

Info & Pricing

Provider

OpenAI

Model Type

Audio

Serverless

Available

Pricing Per 1M Tokens

No Pricing Available