This is a quantized version of distil-whisper-medium.en, optimized with ctranslate2 to use 8-bit integers for faster inference while maintaining accuracy. Ideal for speech-to-text tasks where speed is critical.

Downloads last month
6
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Rejekts/fastest-distil-whisper-medium.en

Finetuned
(7)
this model