Papers
arxiv:2604.25441

Praxy Voice: Voice-Prompt Recovery + BUPS for Commercial-Class Indic TTS from a Frozen Non-Indic Base at Zero Commercial-Training-Data Cost

Published on Apr 28
· Submitted by
Venkata Pushpak Teja Menta
on Apr 30

Abstract

Researchers enhance a non-Indic-native text-to-speech system by implementing a Brahmic Unified Phoneme Space, LoRA adaptation, and voice-prompt recovery techniques to achieve commercial-quality output for Indic languages without requiring new acoustic decoders or commercial training data.

AI-generated summary

Commercial TTS systems produce near-native Indic audio, but the best open-source bases (Chatterbox, Indic Parler-TTS, IndicF5) trail them on measured phonological dimensions, and the most widely adopted multilingual base (Chatterbox, 23 languages) does not even tokenise Telugu or Tamil. We ask: what is the minimum intervention that brings such a non-Indic-native base to commercial-class output on Telugu, Tamil, and Hindi, without training a new acoustic decoder and without any commercial TTS training data? We combine three pieces: (1) BUPS, a Brahmic Unified Phoneme Space that deterministically romanises seven Indic scripts to ISO-15919 so Chatterbox's Latin tokeniser can process them; (2) a LoRA adapter on only the text-token predictor (Chatterbox's t3), trained on ~1,220h of licensed Indic audio with a Hindi-proxy language_id; (3) a voice-prompt recovery recipe -- an 8-11s same-language reference clip plus three sampling overrides (exaggeration 0.7, temperature 0.6, min_p 0.1; "Config B") -- that recovers commercial-class acoustic output with no acoustic-decoder training. On Hindi, the LoRA regresses accuracy and we instead use vanilla Chatterbox + Config B, giving a two-branch deployment. Evaluated on 10-utterance pilot sets with the companion PSP benchmark, Praxy Voice matches or slightly leads commercial baselines: 26.7% retroflex collapse on Telugu (vs Sarvam Bulbul 33.3%), 71% Tamil-zha collapse (vs commercial trio's 86%), 0.025 LLM-WER on Hindi (tied with Cartesia Sonic-3). For intra-sentential code-mix we add a third branch (IndicF5 + native-script transliteration) that drops code-mix LLM-WER from 0.80-0.85 to 0.14-0.27 across Hi/Te/Ta. We release R6 LoRA weights (Apache-2.0), inference code and router (MIT), and a Gradio demo.

Community

Paper author Paper submitter

Indic TTS from a frozen non-Indic base, at zero commercial-training-data cost.

We start with Chatterbox (English-trained, MIT) and ship commercial-class Telugu/Hindi/Tamil TTS using three contributions:

  1. BUPS: a Brahmic Unified Phoneme Space that romanises Devanagari/Telugu/Tamil through a shared ISO-15919 layer, letting one English-trained base address all three scripts.
  2. Voice-prompt recovery + Config-B sampling: a generation-time recipe that reconstructs Indic phonology from any 8–15s reference clip without retraining the base.
  3. R6 LoRA: a 16-rank adapter on T3 transformer attention, fine-tuned on the publicly-licensed IndicVoices subset only.

Headline finding (PSP benchmark, arXiv:2604.25476): Praxy R6 beats Sarvam Bulbul on Telugu retroflex collapse (26.7% vs 33.3%) and outperforms a commercial trio on Tamil-zha (71% vs 86%), while staying within striking distance of ElevenLabs/Cartesia on Hindi LLM-WER. Apache-2.0 weights, MIT code, live HF Space demo.

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2604.25441
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 1

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2604.25441 in a dataset README.md to link it from this page.

Spaces citing this paper 1

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.