Cheatsheet
Text-To-Speech (TTS) Cheatsheet
A comprehensive cheatsheet, provides an overview of the top open-source TTS models, inference libraries, training resources, and more to help you get started with or enhance your TTS projects.
1. Models (Open-Source)
- XTTS-v2: High-quality TTS model with robust voice quality.
- MeloTTS-English: Specialized English voice synthesis with melodic intonation.
- F5-TTS: Produces high-quality speech with voice cloning and customization.
- Bark: High-quality multilingual model supporting varied accents and prosody.
- Parler-tts-mini-v1: Compact model optimized for quick demos and limited resource environments.
Additional Notable Models
- FastSpeech2: Speed-focused, decent quality trade-off.
- VITS: End-to-end TTS offering high fidelity and voice controllability.
- SpeechT5: Unified-modal SpeechT5 framework that explores encoder-decoder pre-training for self-supervised speech/text representation learning.
2. Inference Libraries / Toolkits
- Coqui TTS: Easy-to-use, community-driven TTS toolkit.
- Parler TTS: Inference and training library for high-quality TTS models.
- LitServe: Lightning-fast inference serving library for quick deployments.
- Mozilla TTS: Well-known toolkit with extensive model support and large community.
- Tortoise TTS: High-quality synthesis; slower but excellent results.
Additional Toolkits
- ESPnet TTS: Unified end-to-end speech processing (ASR + TTS).
- NVIDIA NeMo: State-of-the-art models + easy fine-tuning on NVIDIA GPUs.
3. Use Cases
- Voice Assistants & Virtual Agents
- Audiobooks & Podcast Generation
- Accessibility Tools (for visually impaired users)
- Interactive Learning & E-Learning Content
- Customer Support Bots & IVR Systems
- Content Localization & Dubbing for Media
4. Deployment Options
- AWS: Use Amazon Sagemaker to deploy for on-demand deployment.
- Azure: Azure Cognitive Services for scalable TTS APIs.
- Inferless: Simplify serverless model serving with no-infrastructure hosting.
- On-Prem: Deploy models on your organization’s local servers.
- Edge Local: Deploy on local servers, Jetson Nano, or Raspberry Pi.
- Containers & Orchestration: Use Docker, K8s for microservice TTS inference.
5. Datasets
- keithito/lj_speech: Popular single-speaker dataset for English TTS.
- facebook/multilingual_librispeech: Multilingual speech corpus for polyglot models.
- amphion/Emilia-Dataset: Specialized dataset for unique voice profiles.
- speechcolab/gigaspeech: Large-scale English speech corpus.
- parler-tts/mls_eng: English subset of the Multilingual LibriSpeech dataset.
More Datasets
- VCTK: High-quality, multi-speaker dataset for diverse accents.
- Common Voice: Crowdsourced multilingual dataset.
- LibriTTS: Enhanced LibriSpeech variant for better TTS results.
6. Training & Fine-Tuning Resources
- GitHub TTS Notebooks & Tutorials: Community-driven code examples and scripts.
- Hugging Face Audio Course, Unit 6 (From Text to Speech): Hands-on tutorials.
- Fine-Tuning a 🐸 TTS Model (Coqui TTS docs): Step-by-step instructions.
- NVIDIA NeMo TTS Guides: Hands-on TTS tutorial notebooks.
- VITS Fast Fine-tuning: Guide you to add your own character voices, or even your own voice, into existing VITS TTS model.
7. Evaluation & Benchmarking
- Mean Opinion Score (MOS), Comparative MOS (CMOS): Subjective quality assessment.
- PESQ, POLQA, NISQA: Objective speech quality metrics.
8. Model Optimization & Compression
- Quantization: Reduce model size & inference time (ONNX).
- Pruning & Distillation: Tailor models to resource constraints.
- Hardware Acceleration: GPUs, TPUs, or specialized inference chips.
- ONNX / TensorRT: Optimize models for low-latency, high-throughput inference.
9. Integration & Workflow Tools
- Gradio / Streamlit: Rapid prototyping with web UIs.
- Airflow / Prefect: Automate data and training pipelines.
- CI/CD (GitHub Actions): Continuous integration for model updates.
- Hugging Face Spaces: Share and demo TTS models easily.
10. Common Challenges & Troubleshooting
- Accents & Dialects: Use multilingual models or phoneme-based TTS.
- Latency Reduction: Optimize models, batch inference, use GPU acceleration.
- Pronunciation Issues: Text normalization and grapheme-to-phoneme conversion.
- Memory Constraints: Use smaller models or pruning/quantization techniques.
11. Ethical Considerations
- Voice Consent & Licensing: Respect dataset/model licenses.
- Disclosure of Synthetic Speech: Inform users when speech is synthesized.
- Bias & Fairness: Be aware of biases in training data and model outputs.
- Deepfake Risks: Implement safeguards and watermarking.
12. Licensing & Governance
- Check Licenses: (MIT, Apache 2.0, GPL) before commercial use.
- Hugging Face Model Cards: Follow best practices for transparency.
- Data Usage Agreements: Ensure compliance with dataset terms.