SpeechT5: Unified-Modal Encoder-Decoder Pre-Training for Spoken Language Processing Paper • 2110.07205 • Published Oct 14, 2021 • 5
SpeechT5 Collection The SpeechT5 framework consists of a shared seq2seq and six modal-specific (speech/text) pre/post-nets that can address a few audio-related tasks. • 8 items • Updated Jan 8 • 24
reazon-research/reazonspeech-nemo-v2 Automatic Speech Recognition • Updated Feb 13, 2024 • 42.5k • 24
Running on CPU Upgrade 12.6k 12.6k Open LLM Leaderboard 🏆 Track, rank and evaluate open LLMs and chatbots