

Reverb
Overview :
Reverb is an open-source inference codebase for speech recognition and speaker segmentation models, utilizing the WeNet framework for ASR and the Pyannote framework for speaker segmentation. It offers detailed model descriptions and allows users to download models from Hugging Face. Reverb aims to provide developers and researchers with high-quality tools for various speech processing tasks.
Target Users :
The target audience primarily includes researchers, developers, and corporate users in the fields of speech recognition and speaker segmentation. Reverb provides high-quality speech processing tools suitable for tasks such as meeting transcription and phone call analysis.
Use Cases
Automatic speech recognition and speaker segmentation for meeting documentation
Voice content analysis for customer service call recordings
Transcription and speaker identification for courtroom records
Features
Speech recognition code based on the WeNet framework
Speaker segmentation code based on the Pyannote framework
Provides WER and WDER results for long-form speech recognition and speaker segmentation
Supports model downloads via Hugging Face Hub
Offers Docker images to simplify deployment
Compatible with NVIDIA GPUs for enhanced performance
Includes detailed installation and usage instructions
How to Use
1. Ensure that Git Large File Storage (LFS) is installed on your system.
2. Use HUGGINGFACE_ACCESS_TOKEN to download models from the Hugging Face Hub.
3. Clone the Reverb code repository to your local machine.
4. Set up and activate a virtual environment.
5. At the root directory of the code repository, set environment variables to include the ASR directory.
6. Build the Docker image (if required).
7. Run the Docker container (if deploying with Docker).
8. Follow the instructions in README.md for model inference and evaluation.
Featured AI Tools

GPT SoVITS
GPT-SoVITS-WebUI is a powerful zero-shot voice conversion and text-to-speech WebUI. It features zero-shot TTS, few-shot TTS, cross-language support, and a WebUI toolkit. The product supports English, Japanese, and Chinese, providing integrated tools such as voice accompaniment separation, automatic training set splitting, Chinese ASR, and text annotation to help beginners create training datasets and GPT/SoVITS models. Users can experience real-time text-to-speech conversion by inputting a 5-second voice sample, and they can fine-tune the model using only 1 minute of training data to improve voice similarity and naturalness. The product supports environment setup, Python and PyTorch versions, quick installation, manual installation, pre-trained models, dataset formats, pending tasks, and acknowledgments.
AI Speech Synthesis
5.8M

Clone Voice
Clone-Voice is a web-based voice cloning tool that can use any human voice to synthesize speech from text using that voice, or convert one voice to another using that voice. It supports 16 languages including Chinese, English, Japanese, Korean, French, German, and Italian. You can record voice online directly from your microphone. Functions include text-to-speech and voice-to-voice conversion. Its advantages lie in its simplicity, ease of use, no need for N card GPUs, support for multiple languages, and flexible voice recording. The product is currently free to use.
AI Speech Synthesis
3.6M