ComfyUI Extension: TTS Audio Suite

Authored by diodiogod

Created

Updated

207 stars

TTS Audio Suite - Universal multi-engine TTS extension for ComfyUI with unified architecture supporting ChatterBox, F5-TTS, and future engines like RVC. Features modular engine adapters, character voice management, comprehensive SRT subtitle support, and advanced audio processing capabilities.

Custom Nodes (0)

    README

    <a id="readme-top"></a>

    Stargazers Issues Forks Dynamic TOML Badge Ko-Fi

    TTS Audio Suite v4.9.10

    ko-fi

    Universal multi-engine TTS extension for ComfyUI - evolved from the original ChatterBox Voice project.

    <div align="center"> <img src="images/AllNodesShowcase.jpg" alt="TTS Audio Suite Nodes Showcase" /> </div>

    A comprehensive ComfyUI extension providing unified Text-to-Speech and Voice Conversion capabilities through multiple engines including ChatterboxTTS, F5-TTS, Higgs Audio 2, and RVC (Real-time Voice Conversion), with modular architecture designed for extensibility and future engine integrations.

    ๐Ÿš€ Project Evolution Timeline

    ๐ŸŽญ ChatterBox Voice Era               ย ย ย ย ย ย ๐ŸŒŸ Multi-Engine Era
    |                                      ย ย ย ย ย ย ย ย ย ย ย ย ย ย ย ย |        
    v1.0 โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ v1.1 โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ v2.0 โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บย v3.0 โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
    Jun 25          ย ย Jun 25 ย ย ย ย ย ย ย ย Jun 25      ย ย ย ย ย Jul 25        โ”‚
    โ”‚              ย ย ย โ”‚  ย ย ย ย ย ย ย ย ย ย ย ย โ”‚           ย ย ย ย ย โ”‚             โ”‚
    Foundation      ย ย SRTย ย ย ย ย ย ย ย ย ย ย  Modular    ย ย ย ย ย ย F5-TTS +      โ”‚
    ChatterBox      ย ย Subtitlesย ย ย ย ย  Structure  ย  ย ย ย ย Audio         โ”‚
    Voice Cloning   ย ย Timing Nodeย ย ย ย Refactor    ย ย ย ย ย Analyzer      โ”‚
                                                        ย ย ย ย ย ย ย ย ย ย ย ย โ–ผ
    v3.4 โ—„โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€ v3.2 โ—„โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€ v3.1 โ—„โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜
    Jul 25        ย ย ย ย ย ย ย ย  Jul 25        ย ย ย ย ย ย ย ย  Jul 25             
    โ”‚           ย ย ย ย ย ย ย ย    โ”‚            ย ย ย ย ย ย ย ย   โ”‚                  
    Language     ย ย ย ย ย ย ย ย   Pause         ย ย ย ย ย ย ย ย  Character          
    Switching     ย ย ย ย ย ย ย ย  Tags          ย ย ย ย ย ย ย ย  Switching          
    [German:Bob]  ย ย ย ย ย ย ย ย  [pause:1s]    ย ย ย ย ย ย ย ย  [Alice]            
    โ”‚                                                  
    โ”‚         โš™๏ธ TTS Audio Suite Era                                
    โ–ผ         |                                   
    v4.0 โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ v4.3 โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ v4.4 โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ–บ v4.5 โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
    Aug 25           Aug 25           Aug 25         Aug 25         โ”‚
    โ”‚                โ”‚                โ”‚              โ”‚              โ”‚
    BREAKING!        RVC +            Silent         Higgs Audio 2  โ”‚
    Project          Voice            Speech         New TTS Engine โ”‚
    Renamed          Conversion       Analyzer       Voice Cloning  โ”‚
    TTS Audio Suite  + Streaming                                    โ”‚
                                                                    โ–ผ
           v4.9 โ—„โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€ v4.8 โ—„โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€ v4.6 โ—„โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜
           Sep 25            Sep 25                Aug 25
           โ”‚                 โ”‚                     โ”‚
           IndexTTS-2        Chatterbox            VibeVoice
           Emotion           Multilingual          New TTS Engine
           Control           Official (23-lang)    90min Generation
    
    <details> <summary><h2>๐Ÿ“‹ Table of Contents</h2></summary> </details>

    ๐ŸŽฅ Demo Videos

    <div align="center"> <a href="https://youtu.be/aHz1mQ2bvEY"> <img src="https://img.youtube.com/vi/aHz1mQ2bvEY/maxresdefault.jpg" width="400" alt="ChatterBox SRT Voice v3.2 - F5-TTS Integration & Features Overview"> </a> <br> <strong><a href="https://youtu.be/aHz1mQ2bvEY">โ–ถ๏ธ v3.2 Features Overview (20min) - F5-TTS Integration, Speech Editor & More!</a></strong> </div> <br> <div align="center"> <a href="https://youtu.be/VyOawMrCB1g?si=7BubljRhsudGqG3s"> <img src="https://img.youtube.com/vi/VyOawMrCB1g/maxresdefault.jpg" width="400" alt="ChatterBox SRT Voice Demo"> </a> <br> <strong><a href="https://youtu.be/VyOawMrCB1g?si=7BubljRhsudGqG3s">โ–ถ๏ธ Original Demo - SRT Timing & Basic Features</a></strong> </div> <details> <summary><h3>๐Ÿ“œ Original ShmuelRonen ChatterBox TTS Nodes</h3></summary> <div align="center"> <img src="https://github.com/user-attachments/assets/4197818c-8093-4da4-abd5-577943ac902c" width="45%" alt="ChatterBox TTS Nodes" /> <img src="https://github.com/user-attachments/assets/701c219b-12ff-4567-b414-e58560594ffe" width="45%" alt="ChatterBox Voice Capture" /> </div>
    • Voice Recording: Smart silence detection for voice capture
    • Enhanced Chunking: Intelligent text splitting with multiple combination methods
    • Unlimited Text Length: No character limits with smart processing

    Original creator: ShmuelRonen

    </details> <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    Features

    • ๐ŸŽค Multi-Engine TTS - ChatterBox TTS, Chatterbox Multilingual TTS, F5-TTS, Higgs Audio 2, VibeVoice, and IndexTTS-2 with voice cloning, reference audio synthesis, and production-grade quality
    • ๐Ÿ”„ Voice Conversion - ChatterBox VC with iterative refinement + RVC real-time conversion using .pth character models
    • ๐ŸŽ™๏ธ Voice Capture & Recording - Smart silence detection and voice input recording
    • ๐ŸŽญ Character & Language Switching - Multi-character TTS with [CharacterName] tags, alias system, and [language:character] syntax for seamless model switching
    • ๐ŸŒ Multi-language Support - Chatterbox Multilingual TTS (Arabic, Danish, German, Greek, English, Spanish, Finnish, French, Hebrew, Hindi, Italian, Japanese, Korean, Malay, Dutch, Norwegian, Polish, Portuguese, Russian, Swedish, Swahili, Turkish, Chinese) + ChatterBox community models (English, German, Italian, French, Russian, Armenian, Georgian, Japanese, Korean, Norwegian) + F5-TTS (English, German, Spanish, French, Japanese, Hindi, and more)
    • ๐Ÿ˜ค Emotion Control - ChatterBox exaggeration parameter for expressive speech + IndexTTS-2 advanced emotion control with dynamic text analysis, character tags, and 8-emotion vectors โ†’ ๐Ÿ“– IndexTTS-2 Guide
    • ๐Ÿ“ Enhanced Chunking - Intelligent text splitting for long content with multiple combination methods
    • ๐ŸŽต Advanced Audio Processing - Optional FFmpeg support for premium audio quality with graceful fallback
    • ๐Ÿค Vocal/Noise Removal - AI-powered vocal separation, noise reduction, and echo removal with GPU acceleration โ†’ ๐Ÿ“– Complete Guide
    • ๐ŸŒŠ Audio Wave Analyzer - Interactive waveform visualization and precise timing extraction for F5-TTS workflows โ†’ ๐Ÿ“– Complete Guide
    • ๐Ÿ—ฃ๏ธ Silent Speech Analyzer - Video analysis with experimental viseme detection, mouth movement tracking, and base SRT timing generation from silent video using MediaPipe
    • โš™๏ธ Parallel Processing - Configurable worker-based processing via batch_size parameter (Note: sequential processing with batch_size=0 remains optimal for performance)
    <div align="right"><a href="#-table-of-contents">Back to top</a></div> <details> <summary><h2>๐Ÿ†• What's New in my Project? (click to expand)</h2></summary> <details> <summary><h3>๐Ÿ“บ SRT Timing and TTS Node</h3></summary> <img title="" src="images/srt.png" alt="SRT Node Screenshot" width="500" data-align="center">

    The "ChatterBox SRT Voice TTS" node allows TTS generation by processing SRT content (SubRip Subtitle) files, ensuring precise timing and synchronization with your audio.

    Key SRT Features:

    • SRT style Processing: Uses SRT style to generate TTS, aligning audio with subtitle timings
    • smart_natural Timing Mode: Intelligent shifting logic that prevents overlaps and ensures natural speech flow
    • Adjusted_SRT Output: Provides actual timings for generated audio for accurate post-processing
    • Segment-Level Caching: Only regenerates modified segments, significantly speeding up workflows

    For comprehensive technical information, refer to the SRT_IMPLEMENTATION.md file.

    </details> <details> <summary><h3>๐Ÿ†• F5-TTS Integration and ๐Ÿ†• Audio Analyzer</h3></summary> <img title="" src="images/waveanalgif.gif" alt="Audio Wave gif" width="500" data-align="center">
    • F5-TTS Voice Synthesis: High-quality voice cloning with reference audio + text
    • Audio Wave Analyzer: Interactive waveform visualization for precise timing extraction
    • Multi-language Support: English, German, Spanish, French, Japanese models
    • Speech Editing Workflows: Advanced F5-TTS editing capabilities
    </details> <details> <summary><h3>๐Ÿ—ฃ๏ธ Silent Speech Analyzer</h3></summary>

    NEW in v4.4.0: Video analysis and mouth movement detection for silent video processing!

    • Mouth Movement Analysis: Real-time detection of mouth shapes and movements from video
    • Experimental Viseme Classification: Approximate detection of vowels (A, E, I, O, U) and consonants (B, F, M, etc.) - results are experimental approximations, not precise
    • 3-Level Analysis System:
      • Frame-level mouth movement detection
      • Syllable grouping with temporal analysis
      • Word prediction using CMU Pronouncing Dictionary (135K+ words)
    • Base SRT Generation: Creates timing-focused SRT files with start/end speech timing as foundation for user editing
    • MediaPipe Integration: Production-ready analysis using Google's MediaPipe framework
    • Visual Feedback: Preview videos with overlaid detection results
    • Automatic Phonetic Placeholders: Word predictions provide phonetically-sensible placeholders, but phrases require user editing for meaningful content
    • TTS Integration: SRT output designed for use with TTS SRT nodes after manual content editing

    Perfect for:

    • Creating base timing templates from silent video footage
    • Animation and VFX reference timing
    • Foundation for manual subtitle creation

    Important Notes:

    • OpenSeeFace provider is experimental and not recommended for production use - MediaPipe is the stable solution
    • Viseme detection is experimental approximation - expect to manually edit both timing and content
    • Generated text placeholders are phonetic suggestions, not meaningful sentences
    </details> <details> <summary><h3>๐ŸŽ™๏ธ Higgs Audio 2 Voice Cloning</h3></summary>

    NEW in v4.5.0: State-of-the-art voice cloning technology with advanced neural voice replication!

    • High-Quality Voice Cloning: Clone any voice from 30+ second reference audio with exceptional fidelity
    • Multi-Speaker Conversations: Native support for character switching within conversations
    • Real-Time Processing: Generate speech in cloned voices with minimal latency
    • Universal Integration: Works seamlessly with existing TTS Text and TTS SRT nodes

    Key Capabilities:

    • Voice Cloning from Reference Audio: Upload any 30+ second audio file for voice replication
    • Multi-Language Support: English (tested), with potential support for Chinese, Korean, German, and Spanish (based on model training data)
    • Character Switching: Use [CharacterName] syntax for multi-speaker dialogues
    • Advanced Generation Control: Fine-tune temperature, top-p, top-k, and token limits
    • Smart Chunking: Automatic handling of unlimited text length with seamless audio combination
    • Intelligent Caching: Instant regeneration of previously processed content

    Technical Features:

    • Modular Architecture: Clean integration with unified TTS system
    • Automatic Model Management: Downloads and organizes models in ComfyUI/models/TTS/HiggsAudio/ structure
    • Progress Tracking: Real-time generation feedback with tqdm progress bars
    • Voice Reference Discovery: Flexible voice file management system

    Quick Start:

    1. Add Higgs Audio Engine node to configure voice cloning parameters
    2. Connect to TTS Text or TTS SRT node for generation
    3. Specify reference audio file or use voice discovery system
    4. Generate high-quality cloned speech with automatic optimization

    Perfect for:

    • Voice acting and character dialogue creation
    • Audiobook narration with consistent voice characteristics
    • Multi-speaker content with distinct voice personalities
    • Professional voice replication for content creation
    </details> <details> <summary><h3>๐ŸŽต VibeVoice Long-Form Generation</h3></summary>

    NEW in v4.6.0: Microsoft VibeVoice engine for unprecedented long-form audio generation!

    • 90-Minute Generation Capability: Generate up to 90 minutes of continuous audio in a single session
    • Dual Multi-Speaker Modes: Choose between Custom Character Switching and Native Multi-Speaker for different workflow needs
    • Microsoft Quality: Official Microsoft VibeVoice models (1.5B and 7B parameter variants) with production-grade output
    • Advanced Parameter Control: CFG scale, sampling modes, temperature, and token limits for fine-tuned generation

    Key Capabilities:

    • Long-Form Audio: Break through traditional TTS length limitations with 90-minute generation
    • Custom Character Switching: Use [Alice], [Bob] character tags with voice files from the voices folder - supports unlimited characters with pause tags and per-character control
    • Native Multi-Speaker: Efficient single-pass generation supporting both [Character] tag auto-conversion and manual "Speaker 1: Hello" format for up to 4 speakers
    • Voice File Integration: Seamless compatibility with existing voice folder structure and Character Voices node
    • Smart Chunking: Automatic text chunking with configurable time-based limits for memory efficiency
    • Priority System: Connected speaker2/3/4_voice inputs override character aliases with intelligent warnings

    Technical Features:

    • Dual Model Support: Microsoft vibevoice-1.5B (2.7B params, faster) and community vibevoice-7B (9.3B params, higher quality)
    • Intelligent Caching: Advanced caching system with mode-aware invalidation for instant regeneration
    • Memory Optimization: Configurable chunking system balances quality with memory usage
    • Unified Architecture: Seamless integration with existing TTS Text and TTS SRT nodes

    Quick Start:

    1. Add โš™๏ธ VibeVoice Engine node to configure model and multi-speaker mode
    2. Connect to TTS Text or TTS SRT node for generation
    3. Choose between Custom Character Switching (recommended) or Native Multi-Speaker mode
    4. Generate long-form content with automatic voice cloning from your voices folder

    Perfect for:

    • Long-form audiobooks and narration with consistent voice quality
    • Multi-character dialogue and conversations with distinct speaker voices
    • Extended podcast-style content with natural speech patterns
    • Educational content requiring extended generation without quality degradation
    </details> <details> <summary><h3>๐ŸŽญ Character & Narrator Switching</h3></summary>

    NEW in v3.1.0: Seamless character switching for both F5TTS and ChatterBox engines!

    • Multi-Character Support: Use [CharacterName] tags to switch between different voices
    • Voice Folder Integration: Organized character voice management system
    • ๐Ÿท๏ธ Character Aliases: User-friendly alias system - use [Alice] instead of [female_01] with #character_alias_map.txt
    • Robust Fallback: Graceful handling when characters not found (no errors!)
    • Universal Compatibility: Works with both F5TTS and ChatterBox TTS engines
    • SRT Integration: Character switching within subtitle timing
    • Backward Compatible: Existing workflows work unchanged

    ๐Ÿ“– Complete Character Switching Guide

    Example usage:

    Hello! This is the narrator speaking.
    [Alice] Hi there! I'm Alice, nice to meet you.
    [Bob] And I'm Bob! Great to meet you both.
    Back to the narrator for the conclusion.
    
    </details> <details> <summary><h3>๐ŸŒ Language Switching with Bracket Syntax</h3></summary>

    NEW in v3.4.0: Seamless language switching using simple bracket notation!

    • Language Code Syntax: Use [language:character] tags to switch languages and models automatically
    • Smart Model Loading: Automatically loads correct language models (F5-DE, F5-FR, German, Norwegian, etc.)
    • Flexible Aliases (v3.4.3): Support for [German:Alice], [Brazil:Bob], [USA:], [Portugal:] - no need to remember language codes!
    • Standard Format: Also supports traditional [fr:Alice], [de:Bob], or [es:] (language only) patterns
    • Character Integration: Combines perfectly with character switching and alias system
    • Performance Optimized: Language groups processed efficiently to minimize model switching
    • Alias Support: Language defaults work with character alias system

    Supported Languages:

    • F5-TTS: English (en), German (de), Spanish (es), French (fr), Italian (it), Japanese (jp), Thai (th), Portuguese (pt), Hindi (hi)
    • ChatterBox: English (en), German (de, de-best, de-expressive), Italian (it), French (fr), Russian (ru), Armenian (hy), Georgian (ka), Japanese (ja), Korean (ko), Norwegian (no/nb/nn)

    Example usage:

    Hello! This is English text with the default model.
    [de:Alice] Hallo! Ich spreche Deutsch mit Alice's Stimme.
    [fr:] Bonjour! Je parle franรงais avec la voix du narrateur.
    [es:Bob] ยกHola! Soy Bob hablando en espaรฑol.
    Back to English with the original model.
    

    Advanced SRT Integration:

    1
    00:00:01,000 --> 00:00:04,000
    Hello! Welcome to our multilingual show.
    
    2
    00:00:04,500 --> 00:00:08,000
    [de:female_01] Willkommen zu unserer mehrsprachigen Show!
    
    3
    00:00:08,500 --> 00:00:12,000
    [fr:] Bienvenue ร  notre รฉmission multilingue!
    
    </details> <details> <summary><h3>๐Ÿ”„ Iterative Voice Conversion</h3></summary>

    NEW: Progressive voice refinement with intelligent caching for instant experimentation!

    • Refinement Passes: Multiple conversion iterations (1-30, recommended 1-5)
    • Smart Caching: Results cached up to 5 iterations - change from 5โ†’3โ†’4 passes instantly
    • Progressive Quality: Each pass refines output to sound more like target voice
    </details> <details> <summary><h3>๐ŸŽต RVC Voice Conversion Integration</h3></summary>

    NEW in v4.1.0: Professional-grade Real-time Voice Conversion with .pth character models!

    • RVC Character Models: Load .pth voice models with ๐ŸŽญ Load RVC Character Model node
    • Unified Voice Changer: Full RVC integration in the Voice Changer node
    • Iterative Refinement: 1-30 passes with smart caching (like ChatterBox)
    • Enhanced Quality: Automatic .index file loading for improved voice similarity
    • Auto-Download: Required models download from official sources automatically
    • Cache Intelligence: Skip recomputation - change 5โ†’3โ†’4 passes instantly
    • Neural Network Quality: High-quality voice conversion using trained RVC models

    ๐Ÿ“– See RVC Models Setup for detailed installation guide

    How it works:

    1. Load your .pth RVC model with ๐ŸŽญ Load RVC Character Model
    2. Connect to ๐Ÿ”„ Voice Changer, select "RVC" engine
    3. Process with iterative refinement for progressive quality improvement
    4. Results cached for instant experimentation with different pass counts
    </details> <details> <summary><h3>โธ๏ธ Pause Tags System</h3></summary>

    NEW: Intelligent pause insertion for natural speech timing control!

    • Smart Pause Syntax: Use pause tags anywhere in your text with multiple aliases
    • Flexible Duration Formats:
      • Seconds: [pause:1.5], [wait:2s], [stop:3]
      • Milliseconds: [pause:500ms], [wait:1200ms], [stop:800ms]
      • Supported aliases: pause, wait, stop (all work identically)
    • Character Integration: Pause tags work seamlessly with character switching
    • Intelligent Caching: Changing pause durations won't regenerate unchanged text segments
    • Universal Support: Works across all TTS nodes (ChatterBox, F5-TTS, SRT)
    • Automatic Processing: No additional parameters needed - just add tags to your text

    Example usage:

    Welcome to our show! [pause:1s] Today we'll discuss exciting topics.
    [Alice] I'm really excited! [wait:500ms] This will be great.
    [stop:2] Let's get started with the main content.
    
    </details> <details> <summary><h3>๐ŸŒ Multi-language ChatterBox Community Models</h3></summary>

    NEW in v4.6.29: ChatterBox TTS now supports 11 languages with community-finetuned models and automatic model management!

    Supported Languages:

    • ๐Ÿ‡บ๐Ÿ‡ธ English: Original ResembleAI model (default)
    • ๐Ÿ‡ฉ๐Ÿ‡ช German: Three variants available:
      • Standard German (stlohrey/chatterbox_de)
      • German Best (havok2) - Multi-speaker hybrid, best quality
      • German Expressive (SebastianBodza) - Emotion control with <haha>, <wow> tags
    • ๐Ÿ‡ฎ๐Ÿ‡น Italian: Bilingual Italian/English model with [it] prefix for Italian text
    • ๐Ÿ‡ซ๐Ÿ‡ท French: 1,400 hours Emilia dataset with zero-shot voice cloning
    • ๐Ÿ‡ท๐Ÿ‡บ Russian: Complete model with training artifacts
    • ๐Ÿ‡ฆ๐Ÿ‡ฒ Armenian: Complete model with unique architecture
    • ๐Ÿ‡ฌ๐Ÿ‡ช Georgian: Complete model with specialized features
    • ๐Ÿ‡ฏ๐Ÿ‡ต Japanese: Uses shared English components with Japanese text processing
    • ๐Ÿ‡ฐ๐Ÿ‡ท Korean: Uses shared English components with Korean text processing
    • ๐Ÿ‡ณ๐Ÿ‡ด Norwegian: Norwegian ChatterBox model (akhbar/chatterbox-tts-norwegian)

    Key Features:

    • Language Dropdown: Simple language selection in all ChatterBox nodes
    • Auto-Download: Models download automatically on first use (~1GB per language)
    • Local Priority: Prefers locally installed models over downloads for offline use
    • Safetensors Support: Modern format support for newer language models
    • Seamless Integration: Works with existing workflows - just select your language

    Usage: Select language from dropdown โ†’ First generation downloads model โ†’ Subsequent generations use cached model

    </details> <details> <summary><h3>๐ŸŒ Chatterbox Multilingual TTS (Official 23-Lang)</h3></summary>

    NEW in v4.8.0: Official ResembleAI Chatterbox Multilingual TTS model with native support for 23 languages!

    The Chatterbox Multilingual TTS (referred to internally as "ChatterBox Official 23-Lang" to distinguish from community models) is ResembleAI's first production-grade open-source TTS model supporting 23 languages out of the box. This is the official successor to the original ChatterBox model with enhanced multilingual capabilities.

    ๐ŸŽฏ Key Advantages over Community Models:

    • Single Unified Model: One model handles all 23 languages - no model switching required
    • Language Parameter Switching: Changes language via parameter, not model loading (faster)
    • Zero-Shot Voice Cloning: Clone any voice with just a few seconds of reference audio across all languages
    • Production-Grade Quality: Benchmarked against leading closed-source systems like ElevenLabs
    • MIT Licensed: Fully open-source with commercial usage rights
    • Perth Watermarking: Built-in responsible AI usage (disabled by default for compatibility)

    ๐ŸŒ Supported Languages (23 total):

    Arabic (ar), Danish (da), German (de), Greek (el), English (en), Spanish (es), Finnish (fi), French (fr), Hebrew (he), Hindi (hi), Italian (it), Japanese (ja), Korean (ko), Malay (ms), Dutch (nl), Norwegian (no), Polish (pl), Portuguese (pt), Russian (ru), Swedish (sv), Swahili (sw), Turkish (tr), Chinese (zh)

    ๐Ÿ”ง Fully Integrated Features:

    • โœ… Character Switching: Full [CharacterName] support with per-character voice references
    • โœ… Language Switching: [language:character] syntax with intelligent parameter switching
    • โœ… Pause Tags: Complete [pause:Ns] support with character voice inheritance
    • โœ… SRT Processing: Advanced subtitle timing with overlapping modes and audio assembly
    • โœ… Voice Conversion: Built-in VC engine supporting all 23 languages
    • โœ… Emotion Control: Unique exaggeration parameter for expressive speech
    • โœ… Advanced Parameters: Full control over repetition_penalty, min_p, top_p for fine-tuning
    • โœ… Cache Invalidation: Proper parameter-based caching for responsive generation

    ๐Ÿ†š vs Community Models:

    | Feature | Chatterbox Multilingual TTS | Community Models | |---------|----------------------------|------------------| | Languages | 23 native languages | 11 finetuned variants | | Model Loading | Single model, parameter switching | Separate model per language | | Voice Cloning | Zero-shot across all languages | Per-model training | | Official Support | โœ… ResembleAI official | Community maintained | | Character Integration | โœ… Full integration | โœ… Full integration | | SRT Support | โœ… Advanced timing modes | โœ… Advanced timing modes | | Performance | Optimized single-model | Multiple model overhead |

    ๐ŸŽญ Character Example:

    [En:Alice] Hello everyone! [De:Hans] Guten Tag! [Es:Maria] ยกHola! [pause:2s] [En:Alice] That was amazing multilingual switching!
    

    This creates seamless multilingual character switching with proper voice inheritance and pause support - all within a single model.

    </details> <details> <summary><h3>โš™๏ธ Universal Streaming Architecture</h3></summary>

    NEW in v4.3.0: Complete architectural overhaul implementing universal streaming system with parallel processing capabilities!

    Key Features:

    • Universal Streaming Infrastructure: Unified processing system eliminating engine-specific code complexity
    • Parallel Processing: Configurable worker-based processing via batch_size parameter
    • Thread-Safe Design: Stateless wrapper architecture eliminates shared state corruption
    • Future-Proof: New engines require only adapter implementation

    Performance Notes:

    • Sequential Recommended: Use batch_size=0 for optimal performance (sequential processing)
    • Parallel Available: batch_size > 1 enables parallel workers but typically slower due to GPU inference characteristics
    • Memory Efficiency: Improved model sharing prevents memory exhaustion when switching modes

    โ†’ ๐Ÿ“– Read Technical Details

    </details> <details> <summary><h3>๐ŸŒˆ IndexTTS-2 With Emotion Control</h3></summary>

    NEW in v4.9.0: Revolutionary IndexTTS-2 engine with advanced emotion control and unified emotion architecture!

    • Unified Emotion Control: Single emotion_control input supporting multiple emotion methods with intelligent priority system
    • Dynamic Text Emotion: AI-powered QwenEmotion analysis with dynamic {seg} template processing for contextual per-segment emotions
    • Direct Audio Reference: Use any audio file as emotion reference for natural emotional expression
    • Character Voices Integration: Use Character Voices opt_narrator output as emotion reference with automatic detection
    • 8-Emotion Vector Control: Manual precision control over Happy, Angry, Sad, Surprised, Afraid, Disgusted, Calm, and Melancholic emotions
    • Character Tag Emotions: Per-character emotion control using [Character:emotion_ref] syntax (highest priority)
    • Emotion Alpha Control: Fine-tune emotion intensity from 0.0 (neutral) to 2.0 (maximum dramatic expression)

    Key Features:

    • Emotion Priority System: Character tags > Global emotion control with intelligent override handling
    • Dynamic Templates: Use {seg} placeholder for contextual emotion analysis (e.g., "Worried parent speaking: {seg}")
    • Universal Compatibility: Works with existing TTS Text and TTS SRT nodes seamlessly
    • Advanced Caching: Stable audio content hashing for reliable cache hits across sessions
    • QwenEmotion Integration: State-of-the-art text emotion analysis with configurable model selection

    Example Usage:

    Welcome to our show! [Alice:happy_sarah] I'm so excited to be here!
    [Bob:angry_narrator] That's completely unacceptable behavior.
    

    Perfect for:

    • Multi-character dialogue with individual emotional expressions
    • Dynamic storytelling with contextual emotion adaptation
    • Professional voice acting with precise emotional control
    • Content creation requiring sophisticated emotional nuance

    ๐Ÿ“– Complete IndexTTS-2 Emotion Control Guide

    </details> </details>

    ๐Ÿš€ Quick Start

    Option 1: ComfyUI Manager (Recommended) โœจ

    One-click installation with intelligent dependency management:

    1. Use ComfyUI Manager to install "TTS Audio Suite"
    2. That's it! ComfyUI Manager automatically runs our install.py script which handles:
      • โœ… Python 3.13 compatibility (MediaPipe โ†’ OpenSeeFace fallback)
      • โœ… Dependency conflicts (NumPy, librosa, etc.)
      • โœ… All bundled engines (ChatterBox, F5-TTS, Higgs Audio)
      • โœ… RVC voice conversion dependencies
      • โœ… Intelligent conflict resolution with --no-deps handling

    Python 3.13 Support:

    • ๐ŸŸข All TTS engines: ChatterBox, F5-TTS, Higgs Audio โœ… Working
    • ๐ŸŸข RVC voice conversion: โœ… Working
    • ๐ŸŸข OpenSeeFace mouth movement: โœ… Working (experimental)
    • ๐Ÿ”ด MediaPipe mouth movement: โŒ Incompatible (use OpenSeeFace)

    Option 2: Manual Installation

    Same intelligent installer, manual setup:

    1. Clone the repository

      cd ComfyUI/custom_nodes
      git clone https://github.com/diodiogod/TTS-Audio-Suite.git
      cd TTS-Audio-Suite
      
    2. Run the intelligent installer:

      ComfyUI Portable:

      # Windows:
      ..\..\..\python_embeded\python.exe install.py
      
      # Linux/Mac:
      ../../../python_embeded/python.exe install.py
      

      ComfyUI with venv/conda:

      # First activate your ComfyUI environment, then:
      python install.py
      

      The installer automatically handles all dependency conflicts and Python version compatibility.

    3. Manual Download Models (OR It will auto-download on first run)

      • Download from HuggingFace ChatterBox
      • Place in ComfyUI/models/TTS/chatterbox/English/ (recommended) or ComfyUI/models/chatterbox/ (legacy)
    4. Try a Workflow

    5. Restart ComfyUI and look for ๐ŸŽค TTS Audio Suite nodes

    ๐Ÿงช Python 3.13 Users: Installation is fully supported! The system automatically uses OpenSeeFace for mouth movement analysis when MediaPipe is unavailable.

    Need F5-TTS? Also download F5-TTS models to ComfyUI/models/F5-TTS/ from the links in the detailed installation below.

    <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    Installation

    <details> <summary>๐Ÿ“‹ Detailed Installation Guide (Click to expand if you're having dependency issues)</summary>

    This section provides a detailed guide for installing TTS Audio Suite, covering different ComfyUI installation methods.

    Prerequisites

    • ComfyUI installation (Portable, Direct with venv, or through Manager)

    • Python 3.12 or higher

    • System libraries (Linux only):

      # Ubuntu/Debian - Required for audio processing
      sudo apt-get install portaudio19-dev libsamplerate0-dev
      
      # Fedora/RHEL
      sudo dnf install portaudio-devel libsamplerate-devel
      

      ๐Ÿ“‹ Why needed? libsamplerate0-dev provides audio resampling libraries for packages like resampy and soxr. portaudio19-dev enables voice recording features.

    • macOS dependencies:

      brew install portaudio
      
    • Windows: No additional system dependencies needed (libraries come pre-compiled)

    Installation Methods

    1. Portable Installation

    For portable installations, follow these steps:

    1. Clone the repository into the ComfyUI/custom_nodes folder:

      cd ComfyUI/custom_nodes
      git clone https://github.com/diodiogod/TTS-Audio-Suite.git
      
    2. Navigate to the cloned directory:

      cd TTS-Audio-Suite
      
    3. Install the required dependencies. Important: Use the python.exe executable located in your ComfyUI portable installation with environment isolation flags.

      ../../../python_embeded/python.exe -m pip install -r requirements.txt --no-user
      

      Why the --no-user flag?

      • Prevents installing to your system Python's user directory, which can cause import conflicts
      • Ensures packages install only to the portable environment for proper isolation

    2. Direct Installation with venv

    If you have a direct installation with a virtual environment (venv), follow these steps:

    1. Clone the repository into the ComfyUI/custom_nodes folder:

      cd ComfyUI/custom_nodes
      git clone https://github.com/diodiogod/TTS-Audio-Suite.git
      
    2. Activate your ComfyUI virtual environment. This is crucial to ensure dependencies are installed in the correct environment. The method to activate the venv may vary depending on your setup. Here's a common example:

      cd ComfyUI
      . ./venv/bin/activate
      

      or on Windows:

      ComfyUI\venv\Scripts\activate
      
    3. Navigate to the cloned directory:

      cd custom_nodes/TTS-Audio-Suite
      
    4. Install the required dependencies using pip:

      pip install -r requirements.txt
      

    3. Installation through the ComfyUI Manager

    1. Install the ComfyUI Manager if you haven't already.

    2. Use the Manager to install the "TTS Audio Suite" node.

    3. The manager might handle dependencies automatically, but it's still recommended to verify the installation. Navigate to the node's directory:

      cd ComfyUI/custom_nodes/TTS-Audio-Suite
      
    4. Activate your ComfyUI virtual environment (see instructions in "Direct Installation with venv").

    5. If you encounter issues, manually install the dependencies:

      pip install -r requirements.txt
      

    Troubleshooting Dependency Issues

    System Dependencies (Linux)

    Our install script automatically detects missing system libraries and will display helpful error messages like:

    [!] Missing system dependencies detected!
    ============================================================
    SYSTEM DEPENDENCIES REQUIRED
    ============================================================
    โ€ข libsamplerate0-dev (for audio resampling)  
    โ€ข portaudio19-dev (for voice recording)
    
    Please install with:
    # Ubuntu/Debian:
    sudo apt-get install libsamplerate0-dev portaudio19-dev
    
    # Fedora/RHEL:
    sudo dnf install libsamplerate-devel portaudio-devel
    ============================================================
    Then run this install script again.
    

    Python Environment Issues

    A common problem is installing dependencies in the wrong Python environment. Always ensure you are installing dependencies within your ComfyUI's Python environment.

    • Verify your Python environment: After activating your venv or navigating to your portable ComfyUI installation, check the Python executable being used:

      which python
      

      This should point to the Python executable within your ComfyUI installation (e.g., ComfyUI/python_embeded/python.exe or ComfyUI/venv/bin/python).

    • If s3tokenizer fails to install: This dependency can be problematic. Try upgrading your pip and setuptools:

      python -m pip install --upgrade pip setuptools wheel
      

      Then, try installing the requirements again.

    • If you cloned the node manually (without the Manager): Make sure you install the requirements.txt file.

    Updating the Node

    To update the node to the latest version:

    1. Navigate to the node's directory:

      cd ComfyUI/custom_nodes/TTS-Audio-Suite
      
    2. Pull the latest changes from the repository:

      git pull
      
    3. Reinstall the dependencies (in case they have been updated):

      pip install -r requirements.txt
      
    </details>

    1. Clone Repository

    cd ComfyUI/custom_nodes
    git clone https://github.com/diodiogod/TTS-Audio-Suite.git
    

    2. Install Dependencies

    Some dependencies, particularly s3tokenizer, can occasionally cause installation issues on certain Python setups (e.g., Python 3.10, sometimes used by tools like Stability Matrix).

    To minimize potential problems, it's highly recommended to first ensure your core packaging tools are up-to-date in your ComfyUI's virtual environment:

    python -m pip install --upgrade pip setuptools wheel
    

    After running the command above, install the node's specific requirements:

    pip install -r requirements.txt
    

    3. Optional: Install FFmpeg for Enhanced Audio Processing

    ChatterBox Voice now supports FFmpeg for high-quality audio stretching. While not required, it's recommended for the best audio quality:

    Windows:

    winget install FFmpeg
    # or with Chocolatey
    choco install ffmpeg
    

    macOS:

    brew install ffmpeg
    

    Linux:

    # Ubuntu/Debian
    sudo apt-get install ffmpeg
    
    # Fedora
    sudo dnf install ffmpeg
    

    If FFmpeg is not available, ChatterBox will automatically fall back to using the built-in phase vocoder method for audio stretching - your workflows will continue to work without interruption.

    4. Download Models

    Download the ChatterboxTTS models and place them in the new organized structure:

    ComfyUI/models/TTS/chatterbox/    โ† Recommended (new structure)
    

    Or use the legacy location (still supported):

    ComfyUI/models/chatterbox/        โ† Legacy (still works)
    

    Required files:

    • conds.pt (105 KB)
    • s3gen.pt (~1 GB)
    • t3_cfg.pt (~1 GB)
    • tokenizer.json (25 KB)
    • ve.pt (5.5 MB)

    Download from: https://huggingface.co/ResembleAI/chatterbox/tree/main

    4.1. Multilanguage ChatterBox Models (Optional)

    NEW in v3.3.0: ChatterBox now supports multiple languages! Models will auto-download on first use, or you can manually install them for offline use.

    For manual installation, create language-specific folders in the organized structure:

    ComfyUI/models/TTS/chatterbox/    โ† Recommended structure
    โ”œโ”€โ”€ English/          # Optional - for explicit English organization
    โ”‚   โ”œโ”€โ”€ conds.pt
    โ”‚   โ”œโ”€โ”€ s3gen.pt
    โ”‚   โ”œโ”€โ”€ t3_cfg.pt
    โ”‚   โ”œโ”€โ”€ tokenizer.json
    โ”‚   โ””โ”€โ”€ ve.pt
    โ”œโ”€โ”€ German/           # German language models
    โ”‚   โ”œโ”€โ”€ conds.safetensors
    โ”‚   โ”œโ”€โ”€ s3gen.safetensors
    โ”‚   โ”œโ”€โ”€ t3_cfg.safetensors
    โ”‚   โ”œโ”€โ”€ tokenizer.json
    โ”‚   โ””โ”€โ”€ ve.safetensors
    โ””โ”€โ”€ Norwegian/        # Norwegian language models
        โ”œโ”€โ”€ conds.safetensors
        โ”œโ”€โ”€ s3gen.safetensors
        โ”œโ”€โ”€ t3_cfg.safetensors
        โ”œโ”€โ”€ tokenizer.json
        โ””โ”€โ”€ ve.safetensors
    

    Note: Legacy location ComfyUI/models/chatterbox/ still works for backward compatibility.

    Available ChatterBox Language Models:

    | Language | HuggingFace Repository | Format | Auto-Download | | ----------------------- | ----------------------------------------------------------------------------------------- | ------------ | ------------- | | English | ResembleAI/chatterbox | .pt | โœ… | | German | stlohrey/chatterbox_de | .safetensors | โœ… | | German (havok2) | niobures/Chatterbox-TTS | .safetensors | โœ… | | German (SebastianBodza) | niobures/Chatterbox-TTS | .safetensors | โœ… | | Italian | niobures/Chatterbox-TTS | .pt | โœ… | | French | Thomcles/ChatterBox-fr | .safetensors | โœ… | | Russian | niobures/Chatterbox-TTS | .safetensors | โœ… | | Armenian | niobures/Chatterbox-TTS | .safetensors | โœ… | | Georgian | niobures/Chatterbox-TTS | .safetensors | โœ… | | Japanese | niobures/Chatterbox-TTS | .safetensors | โœ… | | Korean | niobures/Chatterbox-TTS | .safetensors | โœ… | | Norwegian | akhbar/chatterbox-tts-norwegian | .safetensors | โœ… |

    Usage: Simply select your desired language from the dropdown in ChatterBox TTS or SRT nodes. First generation will auto-download the model.

    4.2. Chatterbox Multilingual TTS (Official 23-Lang)

    Repository: ResembleAI/chatterbox | Size: ~4.3GB | Auto-Download: โœ…

    The official ResembleAI multilingual model supporting 23 languages in a unified model structure.

    Installation Structure:

    ComfyUI/models/TTS/chatterbox_official_23lang/
    โ””โ”€โ”€ Official 23-Lang/
        โ”œโ”€โ”€ t3_23lang.safetensors    # Multilingual T3 model (2.14GB)
        โ”œโ”€โ”€ s3gen.pt                 # S3Gen model (1.06GB)
        โ”œโ”€โ”€ ve.pt                   # Voice encoder (5.7MB) 
        โ”œโ”€โ”€ mtl_tokenizer.json      # Multilingual tokenizer (68KB)
        โ””โ”€โ”€ conds.pt               # Conditioning (107KB, optional)
    

    Model Selection: Choose "ChatterBox Official 23-Lang" from Unified TTS Engine dropdown.

    5. F5-TTS Models (Optional)

    For F5-TTS voice synthesis capabilities, download F5-TTS models and place them in the organized structure:

    ComfyUI/models/TTS/F5-TTS/       โ† Recommended (new structure)
    

    Or use the legacy location (still supported):

    ComfyUI/models/F5-TTS/           โ† Legacy (still works)
    

    Available F5-TTS Models:

    | Model | Language | Download | Size | | ------------------ | ---------------- | -------------------------------------------------------------------------------- | ------ | | F5TTS_Base | English | HuggingFace | ~1.2GB | | F5TTS_v1_Base | English (v1) | HuggingFace | ~1.2GB | | E2TTS_Base | English (E2-TTS) | HuggingFace | ~1.2GB | | F5-DE | German | HuggingFace | ~1.2GB | | F5-ES | Spanish | HuggingFace | ~1.2GB | | F5-FR | French | HuggingFace | ~1.2GB | | F5-JP | Japanese | HuggingFace | ~1.2GB | | F5-Hindi-Small | Hindi | HuggingFace | ~632MB |

    Vocoder (Optional but Recommended):

    ComfyUI/models/TTS/F5-TTS/vocos/     โ† Recommended
    โ”œโ”€โ”€ config.yaml
    โ”œโ”€โ”€ pytorch_model.bin
    โ””โ”€โ”€ vocab.txt
    

    Legacy location also supported: ComfyUI/models/F5-TTS/vocos/

    Download from: Vocos Mel-24kHz

    Complete Folder Structure (Recommended):

    ComfyUI/models/TTS/F5-TTS/
    โ”œโ”€โ”€ F5TTS_Base/
    โ”‚   โ”œโ”€โ”€ model_1200000.safetensors    โ† Main model file
    โ”‚   โ””โ”€โ”€ vocab.txt                    โ† Vocabulary file
    โ”œโ”€โ”€ vocos/                           โ† For offline vocoder
    โ”‚   โ”œโ”€โ”€ config.yaml
    โ”‚   โ””โ”€โ”€ pytorch_model.bin
    โ””โ”€โ”€ F5TTS_v1_Base/
        โ”œโ”€โ”€ model_1250000.safetensors
        โ””โ”€โ”€ vocab.txt
    

    Required Files for Each Model:

    • model_XXXXXX.safetensors - The main model weights
    • vocab.txt - Vocabulary/tokenizer file (download from same HuggingFace repo)

    Note: F5-TTS uses internal config files, no config.yaml needed. Vocos vocoder doesn't need vocab.txt.

    Note: F5-TTS models and vocoder will auto-download from HuggingFace if not found locally. The first generation may take longer while downloading (~1.2GB per model).

    6. F5-TTS Voice References Setup

    For easy voice reference management, create a dedicated voices folder:

    ComfyUI/models/voices/
    โ”œโ”€โ”€ character1.wav
    โ”œโ”€โ”€ character1.reference.txt โ† Contains: "Hello, I am character one speaking clearly."
    โ”œโ”€โ”€ character1.txt          โ† Contains: "BBC Radio sample, licensed under CC3..."
    โ”œโ”€โ”€ narrator.wav
    โ”œโ”€โ”€ narrator.txt            โ† Contains: "This is the narrator voice for storytelling."
    โ”œโ”€โ”€ my_voice.wav
    โ””โ”€โ”€ my_voice.txt            โ† Contains: "This is my personal voice sample."
    

    Voice Reference Requirements:

    • Audio files: WAV format, 5-30 seconds, clean speech, 24kHz recommended
    • Text files: Exact transcription of what's spoken in the audio file
    • Naming: filename.wav + filename.reference.txt (preferred) or filename.txt (fallback)
    • Character Names: Character name = audio filename (without extension). Subfolders supported for organization.

    โš ๏ธ F5-TTS Best Practices: Follow these guidelines to avoid inference failures

    <details> <summary><strong>๐Ÿ“‹ F5-TTS Inference Guidelines</strong></summary>

    To avoid possible inference failures, make sure you follow these F5-TTS optimization guidelines:

    1. Reference Audio Duration: Use reference audio <12s and leave proper silence space (e.g. 1s) at the end. Otherwise there is a risk of truncating in the middle of word, leading to suboptimal generation.

    2. Letter Case Handling: Uppercased letters (best with form like K.F.C.) will be uttered letter by letter, and lowercased letters used for common words.

    3. Pause Control: Add some spaces (blank: " ") or punctuations (e.g. "," ".") to explicitly introduce some pauses.

    4. Punctuation Spacing: If English punctuation marks the end of a sentence, make sure there is a space " " after it. Otherwise not regarded as sentence chunk.

    5. Number Processing: Preprocess numbers to Chinese letters if you want to have them read in Chinese, otherwise they will be read in English.

    These guidelines help ensure optimal F5-TTS generation quality and prevent common audio artifacts.

    </details>

    7. Higgs Audio 2 Models (Optional - NEW in v4.5.0+)

    For state-of-the-art voice cloning capabilities, Higgs Audio 2 models are automatically downloaded to the organized structure:

    ComfyUI/models/TTS/HiggsAudio/        โ† Recommended (new structure)
    โ”œโ”€โ”€ higgs-audio-v2-3B/               โ† Main model directory
    โ”‚   โ”œโ”€โ”€ generation/                  โ† Generation model files
    โ”‚   โ”‚   โ”œโ”€โ”€ config.json
    โ”‚   โ”‚   โ”œโ”€โ”€ model.safetensors.index.json
    โ”‚   โ”‚   โ”œโ”€โ”€ model-00001-of-00003.safetensors (~3GB)
    โ”‚   โ”‚   โ”œโ”€โ”€ model-00002-of-00003.safetensors (~3GB) 
    โ”‚   โ”‚   โ”œโ”€โ”€ model-00003-of-00003.safetensors (~3GB)
    โ”‚   โ”‚   โ”œโ”€โ”€ generation_config.json
    โ”‚   โ”‚   โ”œโ”€โ”€ tokenizer.json
    โ”‚   โ”‚   โ”œโ”€โ”€ tokenizer_config.json
    โ”‚   โ”‚   โ””โ”€โ”€ special_tokens_map.json
    โ”‚   โ””โ”€โ”€ tokenizer/                   โ† Audio tokenizer files
    โ”‚       โ”œโ”€โ”€ config.json
    โ”‚       โ””โ”€โ”€ model.pth (~200MB)
    โ””โ”€โ”€ voices/                          โ† Voice reference files
        โ”œโ”€โ”€ character1.wav               โ† 30+ second reference audio
        โ”œโ”€โ”€ character1.txt               โ† Exact transcription
        โ”œโ”€โ”€ narrator.wav
        โ””โ”€โ”€ narrator.txt
    

    Available Higgs Audio Models (Auto-Download):

    | Model | Type | Source | Size | Auto-Download | | ----------------- | ------------- | ------------------------------------------------------------------------------------------------------------- | ------ | ------------- | | higgs-audio-v2-3B | Voice Cloning | bosonai/higgs-audio-v2-generation-3B-base | ~9GB | โœ… | | Audio Tokenizer | Tokenization | bosonai/higgs-audio-v2-tokenizer | ~200MB | โœ… |

    Voice Reference Requirements:

    • Audio files: WAV format, 30+ seconds, clean speech, single speaker
    • Text files: Exact transcription of the reference audio
    • Naming: filename.wav + filename.txt (transcription)
    • Quality: Clear, noise-free audio for best voice cloning results

    How Higgs Audio Auto-Download Works:

    1. Select Model: Choose "higgs-audio-v2-3B" in Higgs Audio Engine node
    2. Auto-Download: Both generation model (~9GB) and tokenizer (~200MB) download automatically
    3. Voice References: Place reference audio and transcriptions in voices/ folder
    4. Local Cache: Once downloaded, models are used from local cache for fast loading

    Manual Installation (Optional):

    To pre-download models for offline use:

    # Download generation model files to:
    # ComfyUI/models/TTS/HiggsAudio/higgs-audio-v2-3B/generation/
    
    # Download tokenizer files to:  
    # ComfyUI/models/TTS/HiggsAudio/higgs-audio-v2-3B/tokenizer/
    

    Usage: Simply use the โš™๏ธ Higgs Audio 2 Engine node โ†’ Select model โ†’ All required files download automatically!

    8. VibeVoice Models (NEW in v4.6.0+)

    For Microsoft VibeVoice Long-Form TTS, models are automatically downloaded to:

    ComfyUI/models/TTS/VibeVoice/        โ† Recommended (new structure)
    โ”œโ”€โ”€ vibevoice-1.5B/                  โ† Microsoft official model (2.7B params)
    โ”‚   โ”œโ”€โ”€ model-00001-of-00003.safetensors (~2GB)
    โ”‚   โ”œโ”€โ”€ model-00002-of-00003.safetensors (~2GB)
    โ”‚   โ”œโ”€โ”€ model-00003-of-00003.safetensors (~2GB)
    โ”‚   โ”œโ”€โ”€ model.safetensors.index.json
    โ”‚   โ”œโ”€โ”€ config.json
    โ”‚   โ””โ”€โ”€ preprocessor_config.json
    โ””โ”€โ”€ vibevoice-7B/                    โ† Community preview model (9.3B params, 18GB)
        โ”œโ”€โ”€ model-00001-of-00010.safetensors (~1.9GB)
        โ”œโ”€โ”€ model-00002-of-00010.safetensors (~1.9GB)
        โ”œโ”€โ”€ ... (10 model files total)
        โ”œโ”€โ”€ model-00010-of-00010.safetensors (~1.7GB)
        โ”œโ”€โ”€ model.safetensors.index.json
        โ”œโ”€โ”€ config.json
        โ””โ”€โ”€ preprocessor_config.json
    

    Available VibeVoice Models (Auto-Download):

    | Model | Description | Size | Source | Auto-Download | | -------------- | ----------------------------------------------- | ------ | ----------------------------------------------------------------------------- | ------------- | | vibevoice-1.5B | Microsoft official (2.7B params, faster) | ~5.4GB | microsoft/VibeVoice-1.5B | โœ… | | vibevoice-7B | Official Microsoft model (9.3B params, quality) | ~18GB | microsoft/VibeVoice-Large | โœ… |

    Key Features:

    • 90-Minute Generation: Both models support up to 90 minutes of continuous audio
    • Multi-Speaker Support: Native dual-mode multi-speaker generation
    • Voice Cloning: Works with existing voices folder structure for character switching
    • Automatic Management: Models download automatically on first use

    Usage: Simply use the โš™๏ธ VibeVoice Engine node โ†’ Select model โ†’ Connect to TTS Text/SRT โ†’ All files download automatically!

    9. RVC Models (Optional - NEW in v4.0.0+)

    For Real-time Voice Conversion capabilities, RVC models are automatically downloaded to the organized structure:

    ComfyUI/models/TTS/RVC/          โ† Recommended (new structure)
    โ”œโ”€โ”€ Claire.pth                   โ† Character voice models
    โ”œโ”€โ”€ Sayano.pth
    โ”œโ”€โ”€ Mae_v2.pth
    โ”œโ”€โ”€ Fuji.pth
    โ”œโ”€โ”€ Monika.pth
    โ”œโ”€โ”€ content-vec-best.safetensors โ† Base models (auto-download)
    โ”œโ”€โ”€ rmvpe.pt
    โ”œโ”€โ”€ hubert/                      โ† HuBERT models (auto-organized)
    โ”‚   โ”œโ”€โ”€ hubert-base-rvc.safetensors
    โ”‚   โ”œโ”€โ”€ hubert-soft-japanese.safetensors
    โ”‚   โ””โ”€โ”€ hubert-soft-korean.safetensors
    โ””โ”€โ”€ .index/                      โ† Index files for better similarity
        โ”œโ”€โ”€ added_IVF1063_Flat_nprobe_1_Sayano_v2.index
        โ”œโ”€โ”€ added_IVF985_Flat_nprobe_1_Fuji_v2.index
        โ”œโ”€โ”€ Monika_v2_40k.index
        โ””โ”€โ”€ Sayano_v2_40k.index
    

    Note: Legacy location ComfyUI/models/RVC/ still works for backward compatibility.

    Available RVC Character Models (Auto-Download):

    | Model | Type | Source | Auto-Download | | ---------- | --------- | -------------------------------------------------------------------------- | ------------- | | Claire.pth | Character | SayanoAI RVC-Studio | โœ… | | Sayano.pth | Character | SayanoAI RVC-Studio | โœ… | | Mae_v2.pth | Character | SayanoAI RVC-Studio | โœ… | | Fuji.pth | Character | SayanoAI RVC-Studio | โœ… | | Monika.pth | Character | SayanoAI RVC-Studio | โœ… |

    Required Base Models (Auto-Download):

    | Model | Purpose | Source | Size | | ---------------------------- | ---------------- | --------------------------------------------------------------------------------- | ------ | | content-vec-best.safetensors | Voice features | lengyue233/content-vec-best | ~300MB | | rmvpe.pt | Pitch extraction | lj1995/VoiceConversionWebUI | ~55MB |

    How RVC Auto-Download Works:

    1. Select Character Model: Choose from available models in ๐ŸŽญ Load RVC Character Model node
    2. Auto-Download: Models download automatically when first selected (with auto_download=True)
    3. Base Models: Required base models download automatically when RVC engine first runs
    4. Index Files: Optional FAISS index files download for improved voice similarity
    5. Local Cache: Once downloaded, models are used from local cache for fast loading

    UVR Models for Vocal Separation (Auto-Download):

    Additional models for the ๐Ÿค Noise or Vocal Removal node download to ComfyUI/models/TTS/UVR/ (recommended) or ComfyUI/models/UVR/ (legacy) as needed.

    Usage: Simply use the ๐ŸŽญ Load RVC Character Model node โ†’ Select a character โ†’ Connect to Voice Changer node. All required models download automatically!

    10. IndexTTS-2 Models (NEW in v4.9.0+)

    For IndexTTS-2 advanced emotion control capabilities, models are automatically downloaded to the organized structure:

    ComfyUI/models/TTS/IndexTTS/
    โ”œโ”€โ”€ IndexTTS-2/                         โ† Main TTS model
    โ”‚   โ”œโ”€โ”€ config.yaml                     โ† Model configuration
    โ”‚   โ”œโ”€โ”€ feat1.pt                        โ† Feature extraction model 1
    โ”‚   โ”œโ”€โ”€ feat2.pt                        โ† Feature extraction model 2
    โ”‚   โ”œโ”€โ”€ gpt.pth                         โ† GPT model weights
    โ”‚   โ”œโ”€โ”€ s2mel.pth                       โ† Speech-to-mel conversion
    โ”‚   โ”œโ”€โ”€ bpe.model                       โ† BPE tokenizer model
    โ”‚   โ”œโ”€โ”€ wav2vec2bert_stats.pt           โ† Wav2Vec2BERT statistics
    โ”‚   โ””โ”€โ”€ qwen0.6bemo4-merge/             โ† QwenEmotion model (for text emotion)
    โ”‚       โ”œโ”€โ”€ config.json
    โ”‚       โ”œโ”€โ”€ generation_config.json
    โ”‚       โ”œโ”€โ”€ model.safetensors
    โ”‚       โ”œโ”€โ”€ tokenizer.json
    โ”‚       โ”œโ”€โ”€ tokenizer_config.json
    โ”‚       โ””โ”€โ”€ other model files...
    โ””โ”€โ”€ w2v-bert-2.0/                       โ† W2V-BERT semantic features
        โ”œโ”€โ”€ config.json
        โ”œโ”€โ”€ model.safetensors               โ† ~2GB model weights
        โ””โ”€โ”€ preprocessor_config.json
    

    Available IndexTTS-2 Models (Auto-Download):

    | Model | Description | Source | Size | Auto-Download | | -------------------- | ------------------------------------- | ---------------------------------------------------------------------- | --------------- | ------------- | | IndexTTS-2 | Main TTS engine with emotion control | IndexTeam/IndexTTS-2 | Multiple files | โœ… | | w2v-bert-2.0 | W2V-BERT semantic feature extractor | facebook/w2v-bert-2.0 | ~2GB | โœ… | | qwen0.6bemo4-merge | QwenEmotion text analysis (built-in) | Included with IndexTTS-2 | Part of main | โœ… |

    Key Features:

    • Unified Emotion Control: Single emotion_control input supporting multiple emotion sources
    • Dynamic Text Emotion: AI-powered QwenEmotion analysis with {seg} template processing
    • Character Tag Emotions: Per-character emotion control using [Character:emotion_ref] syntax
    • Audio Reference Emotion: Use any audio file or Character Voices as emotion reference
    • 8-Emotion Vector Control: Manual precision control over emotional expression

    Usage: Simply use the โš™๏ธ IndexTTS-2 Engine node โ†’ Connect emotion control โ†’ All required models download automatically!

    11. Restart ComfyUI

    <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    ๐Ÿ’– Support This Project

    If TTS Audio Suite has been helpful for your projects, consider supporting its development:

    ko-fi

    Your support helps maintain and improve this project for the entire community!

    <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    Enhanced Features

    ๐Ÿ“ Intelligent Text Chunking (NEW!)

    Long text support with smart processing:

    • Character-based limits (100-1000 chars per chunk)
    • Sentence boundary preservation - won't cut mid-sentence
    • Multiple combination methods:
      • auto - Smart selection based on text length
      • concatenate - Simple joining
      • silence_padding - Add configurable silence between chunks
      • crossfade - Smooth audio blending
    • Comma-based splitting for very long sentences
    • Backward compatible - works with existing workflows

    Chunking Controls (all optional):

    • enable_chunking - Enable/disable smart chunking (default: True)
    • max_chars_per_chunk - Chunk size limit (default: 400)
    • chunk_combination_method - How to join audio (default: auto)
    • silence_between_chunks_ms - Silence duration (default: 100ms)

    Auto-selection logic:

    • Text > 1000 chars โ†’ silence_padding (natural pauses)
    • Text > 500 chars โ†’ crossfade (smooth blending)
    • Text < 500 chars โ†’ concatenate (simple joining)

    ๐Ÿ“ฆ Smart Model Loading

    Priority-based model detection:

    1. Bundled models in node folder (self-contained)
    2. ComfyUI models in standard location
    3. HuggingFace download with authentication

    Console output shows source:

    ๐Ÿ“ฆ Using BUNDLED ChatterBox (self-contained)
    ๐Ÿ“ฆ Loading from bundled models: ./models/chatterbox
    โœ… ChatterboxTTS model loaded from bundled!
    
    <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    Usage

    Voice Recording

    1. Add "๐ŸŽค ChatterBox Voice Capture" node
    2. Select your microphone from the dropdown
    3. Adjust recording settings:
      • Silence Threshold: How quiet to consider "silence" (0.001-0.1)
      • Silence Duration: How long to wait before stopping (0.5-5.0 seconds)
      • Sample Rate: Audio quality (8000-96000 Hz, default 44100)
    4. Change the Trigger value to start a new recording
    5. Connect output to TTS (for voice cloning) or VC nodes

    Enhanced Text-to-Speech

    1. Add "๐ŸŽค ChatterBox Voice TTS" node
    2. Enter your text (any length - automatic chunking)
    3. Optionally connect reference audio for voice cloning
    4. Adjust TTS settings:
      • Exaggeration: Emotion intensity (0.25-2.0)
      • Temperature: Randomness (0.05-5.0)
      • CFG Weight: Guidance strength (0.0-1.0)

    F5-TTS Voice Synthesis

    1. Add "๐ŸŽค F5-TTS Voice Generation" node
    2. Enter your target text (any length - automatic chunking)
    3. Required: Connect reference audio for voice cloning
    4. Required: Enter reference text that matches the reference audio exactly
    <details> <summary>๐Ÿ“– Voice Reference Setup Options</summary>

    Two ways to provide voice references:

    1. Easy Method: Select voice from reference_audio_file dropdown โ†’ text auto-detected from companion .txt file
    2. Manual Method: Set reference_audio_file to "none" โ†’ connect opt_reference_audio + opt_reference_text inputs
    </details>
    1. Select F5-TTS model:
      • F5TTS_Base: English base model (recommended)
      • F5TTS_v1_Base: English v1 model
      • E2TTS_Base: E2-TTS model
      • F5-DE: German model
      • F5-ES: Spanish model
      • F5-FR: French model
      • F5-JP: Japanese model
    2. Adjust F5-TTS settings:
      • Temperature: Voice variation (0.1-2.0, default: 0.8)
      • Speed: Speech speed (0.5-2.0, default: 1.0)
      • CFG Strength: Guidance strength (0.0-10.0, default: 2.0)
      • NFE Step: Quality vs speed (1-100, default: 32)

    Voice Conversion with Iterative Refinement

    1. Add "๐Ÿ”„ ChatterBox Voice Conversion" node
    2. Connect source audio (voice to convert)
    3. Connect target audio (voice style to copy)
    4. Configure refinement settings:
      • Refinement Passes: Number of conversion iterations (1-30, recommended 1-5)
      • Each pass refines the output to sound more like the target
      • Smart Caching: Results cached up to 5 iterations for instant experimentation

    ๐Ÿง  Intelligent Caching Examples:

    • Run 3 passes โ†’ caches iterations 1, 2, 3
    • Change to 5 passes โ†’ resumes from cached 3, runs 4, 5
    • Change to 2 passes โ†’ returns cached iteration 2 instantly
    • Change to 4 passes โ†’ resumes from cached 3, runs 4

    ๐Ÿ’ก Pro Tip: Start with 1 pass, then experiment with 2-5 passes to find the sweet spot for your audio. Each iteration can improves voice similarity!

    <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    ๐Ÿ“ Example Workflows

    Ready-to-use ComfyUI workflows - Download and drag into ComfyUI:

    ๐Ÿ†• Unified Workflows (v4.5+)

    | Workflow | Description | Features | Status | Files | | ---------------------------- | --------------------------------------------- | -------------------------------------------------------------------------------------------------------------------- | ---------------------- | ------------------------------------------------------------------------------------------- | | Unified ๐Ÿ“บ TTS SRT | Universal SRT processing with all TTS engines | โ€ข ChatterBox/F5-TTS/Higgs Audio 2<br>โ€ข Multiple timing modes<br>โ€ข Multi-character switching<br>โ€ข Overlap SRT support | โœ… New in v4.5 | ๐Ÿ“ JSON | | Unified ๐Ÿ”„ Voice Changer | Modern voice conversion with multiple engines | โ€ข RVC + ChatterBox VC<br>โ€ข Iterative refinement<br>โ€ข Real-time conversion | โœ… Updated for v4.3 | ๐Ÿ“ JSON |

    Specific Workflows

    | Workflow | Description | Status | Files | | -------------------------- | ------------------------------------------------ | -------------------- | -------------------------------------------------------------------------- | | ๐ŸŒˆ IndexTTS-2 Integration | IndexTTS-2 engine with advanced emotion control | โœ… New in v4.9 | ๐Ÿ“ JSON | | VibeVoice Integration | VibeVoice long-form TTS with multi-speaker support | โœ… Compatible | ๐Ÿ“ JSON | | ChatterBox Integration | General ChatterBox TTS and Voice Conversion | โœ… Compatible | ๐Ÿ“ JSON | | F5-TTS Speech Editor | Interactive waveform analysis for F5-TTS editing | โœ… Updated for v4 | ๐Ÿ“ JSON |

    ๐Ÿ’ก Recommended: Use the new Unified ๐Ÿ“บ TTS SRT workflow which showcases all engines and features in one comprehensive workflow. It demonstrates SRT processing, timing modes, multi-character switching, and supports ChatterBox, F5-TTS, and Higgs Audio 2 engines.

    ๐Ÿ“ฅ Usage: Download the .json files and drag them directly into your ComfyUI interface. The workflows will automatically load with proper node connections.

    <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    Settings Guide

    Enhanced Chunking Settings

    For Long Articles/Books:

    • max_chars_per_chunk=600, combination_method=silence_padding, silence_between_chunks_ms=200

    For Natural Speech:

    • max_chars_per_chunk=400, combination_method=auto (default - works well)

    For Fast Processing:

    • max_chars_per_chunk=800, combination_method=concatenate

    For Smooth Audio:

    • max_chars_per_chunk=300, combination_method=crossfade

    Voice Recording Settings

    General Recording:

    • silence_threshold=0.01, silence_duration=2.0 (default settings)

    Noisy Environment:

    • Higher silence_threshold (~0.05) to ignore background noise
    • Longer silence_duration (~3.0) to avoid cutting off speech

    Quiet Environment:

    • Lower silence_threshold (~0.005) for sensitive detection
    • Shorter silence_duration (~1.0) for quick stopping

    TTS Settings

    General Use:

    • exaggeration=0.5, cfg_weight=0.5 (default settings work well)

    Expressive Speech:

    • Lower cfg_weight (~0.3) + higher exaggeration (~0.7)
    • Higher exaggeration speeds up speech; lower CFG slows it down
    <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    Text Processing Capabilities

    ๐Ÿ“š No Hard Text Limits!

    Unlike many TTS systems:

    • OpenAI TTS: 4096 character limit
    • ElevenLabs: 2500 character limit
    • ChatterBox: No documented limits + intelligent chunking

    ๐Ÿง  Smart Text Splitting

    Sentence Boundary Detection:

    • Splits on .!? with proper spacing
    • Preserves sentence integrity
    • Handles abbreviations and edge cases

    Long Sentence Handling:

    • Splits on commas when sentences are too long
    • Maintains natural speech patterns
    • Falls back to character limits only when necessary
    <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    License

    MIT License - Same as ChatterboxTTS

    <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    Credits

    • ResembleAI for ChatterboxTTS
    • ComfyUI team for the amazing framework
    • sounddevice library for audio recording functionality
    • ShmuelRonen for the Original ChatteBox Voice TTS node
    • Diogod for the TTS Audio Suite universal multi-engine implementation
    <div align="right"><a href="#-table-of-contents">Back to top</a></div>

    ๐Ÿ”— Links


    Note: The original ChatterBox model includes Resemble AI's Perth watermarking system for responsible AI usage. This ComfyUI integration includes the Perth dependency but has watermarking disabled by default to ensure maximum compatibility. Users can re-enable watermarking by modifying the code if needed, while maintaining the full quality and capabilities of the underlying TTS model.

    <!-- MARKDOWN LINKS & IMAGES --> <!-- https://www.markdownguide.org/basic-syntax/#reference-style-links -->