Capability
Voice Model Selection And Voice Identity Consistency
20 artifacts provide this capability.
Want a personalized recommendation?
Find the best match →Top Matches
via “multi-speaker voice synthesis from single vits model”
Fast local neural TTS optimized for Raspberry Pi and edge devices.
Unique: Stores speaker mappings in voice configuration JSON rather than requiring separate model files per speaker, enabling efficient multi-voice synthesis with single ONNX model load and minimal memory overhead
vs others: More efficient than loading separate TTS models per voice (e.g., multiple Tacotron2 models); speaker conditioning at inference time adds negligible latency vs. voice switching overhead in alternatives