An API service based on Faster-Whisper for audio transcription using OpenAI's Whisper model.
- REST API for audio file transcription
- CUDA support for GPU acceleration
- Containerized with Docker for easy deployment
- Error handling and reconnection attempts
- Docker
- NVIDIA GPU with CUDA support (optional, but recommended for better performance)
- NVIDIA Container Toolkit (to use GPU in Docker)
docker-compose build --no-cache
docker-compose up -d --force-recreatepython examples/simple_client.py path/to/your/audio.wavSend a POST request to the /transcribe endpoint with an audio file:
import requests
url = "http://localhost:8421/transcribe"
files = {'file': open('audio.wav', 'rb')}
response = requests.post(url, files=files)
print(response.json())The service uses port 8421 by default. You can modify this configuration in the server.py file.