Transcriber Models
Base and fine-tuned transcription models
Whisper Large v3
Base
Active
WER18.4%
CER5.2%
Latency1.2s
Size1.55 GB
Whisper Large v3 — Fine-tuned v2
Fine-tuned
Draft
WER11.2%
CER3.1%
Latency1.3s
Size1.58 GB
Whisper Medium
Base
Archived
WER22.1%
CER7.4%
Latency0.8s
Size769 MB
Whisper Large v3 — Fine-tuned v1
Fine-tuned
Archived
WER14.8%
CER4.3%
Latency1.3s
Size1.57 GB
Model Comparison
Compare all available transcription models across key performance metrics.
| Model | Type | Language | WER % | CER % | Latency | Size | Status |
|---|---|---|---|---|---|---|---|
| Whisper Large v3 | Base | English, Lithuanian | 18.4% | 5.2% | 1.2s | 1.55 GB | Active |
| Whisper Large v3 — Fine-tuned v2 | Fine-tuned | Lithuanian | 11.2% | 3.1% | 1.3s | 1.58 GB | Draft |
| Whisper Medium | Base | English | 22.1% | 7.4% | 0.8s | 769 MB | Archived |
| Whisper Large v3 — Fine-tuned v1 | Fine-tuned | Lithuanian | 14.8% | 4.3% | 1.3s | 1.57 GB | Archived |
WER Benchmarking
Compare model performance on a standard test dataset. Lower WER/CER is better. Latency measured as real-time factor (RTF).
| Model | WER % | CER % | Latency (RTF) | Samples | Status |
|---|---|---|---|---|---|
| Whisper Large v3 — Fine-tuned v2 | 11.2% | 3.1% | 0.42x | 50 / 50 | Complete |
| Whisper Large v3 — Fine-tuned v1 | 14.8% | 4.3% | 0.43x | 50 / 50 | Complete |
| Whisper Large v3 (Base) | 18.4% | 5.2% | 0.38x | 50 / 50 | Complete |
| Whisper Medium (Base) | 22.1% | 7.4% | 0.21x | 50 / 50 | Complete |
WER Comparison
Bar width represents relative WER (lower is better). Best model highlighted in green.