Search:
Match:
2 results

Analysis

This paper addresses the challenge of speech synthesis for the endangered Manchu language, which faces data scarcity and complex agglutination. The proposed ManchuTTS model introduces innovative techniques like a hierarchical text representation, cross-modal attention, flow-matching Transformer, and hierarchical contrastive loss to overcome these challenges. The creation of a dedicated dataset and data augmentation further contribute to the model's effectiveness. The results, including a high MOS score and significant improvements in agglutinative word pronunciation and prosodic naturalness, demonstrate the paper's significant contribution to the field of low-resource speech synthesis and language preservation.
Reference

ManchuTTS attains a MOS of 4.52 using a 5.2-hour training subset...outperforming all baseline models by a notable margin.

Research#TTS🔬 ResearchAnalyzed: Jan 10, 2026 14:15

Scaling TTS LLMs: Multi-Reward GRPO for Enhanced Stability and Prosody

Published:Nov 26, 2025 10:50
1 min read
ArXiv

Analysis

This ArXiv paper explores improvements in text-to-speech (TTS) Large Language Models (LLMs), focusing on stability and prosodic quality. The use of Multi-Reward GRPO suggests a novel approach to training these models, potentially impacting the generation of more natural-sounding speech.
Reference

The research focuses on single-codebook TTS LLMs.