VoxCPMEx
Elixir wrapper for VoxCPM2 — a tokenizer-free, diffusion autoregressive Text-to-Speech model from OpenBMB.
2B parameters · 30 languages · 48kHz output · trained on 2M+ hours of speech data.
⚡ v0.2.0 — MessagePack Protocol + True Streaming
VoxCPMEx v0.2.0 replaces JSON+Base64 with MessagePack binary framing:
- Audio is sent as raw bytes — no base64 encoding, ~33% smaller on the wire
- Control messages are compact msgpack maps
-
Frame format:
[4B BE length][msgpack payload]— no line splitting, no JSON parsing - True streaming:
generate_streaming_async/3returns immediately, poll withnext_chunk/2
Features
- 🌍 30 Languages — Chinese, English, Japanese, Korean, Arabic, French, German, Spanish, Italian, Portuguese, Dutch, Russian, Thai, Vietnamese, Hindi, and 15+ more. Chinese dialects: 四川话, 粤语, 吴语, 东北话, 河南话, 陕西话, 山东话, 天津话, 闽南话
- 🎨 Voice Design — Generate a novel voice from a natural-language description alone. No reference audio needed.
- 🎛️ Controllable Cloning — Clone any voice from a short clip, with optional style guidance (emotion, pace, tone)
- 🎙️ Ultimate Cloning — Continuation cloning with transcript for maximum fidelity
- 🔊 48kHz Studio Output — AudioVAE V2 built-in super-resolution (16kHz in → 48kHz out)
- ⚡ Real-Time Streaming — RTF ~0.3 on RTX 4090, ~0.13 with Nano-VLLM
- 🎓 LoRA Fine-Tuning — Adapt with as little as 5–10 minutes of audio
- 📜 Apache-2.0 — Free for commercial use
Architecture
VoxCPMEx uses Erlang Ports to communicate with a Python process running the VoxCPM2 model. Each VoxCPMEx.start_link/1 spawns a dedicated Python process, allowing multiple models or instances to run concurrently.
+---------------+ msgpack frames +-----------------+
| Elixir | -----------------> | Python |
| GenServer | | VoxCPM2 |
| | <----------------- | |
+---------------+ msgpack frames +-----------------+
raw audio bytes
## Requirements
- Python ≥ 3.10
- PyTorch ≥ 2.5.0, CUDA ≥ 12.0 (or Apple Silicon / CPU)
- Elixir ≥ 1.14
- ~8 GB VRAM (GPU recommended)
## Installation
### 1. Add the dependency
def deps do [
{:voxcpmex, "~> 0.1.0"}] end
### 2. Install Python dependencies
CUDA (NVIDIA GPU) — recommended
mix voxcpmex.setup
Apple Silicon
mix voxcpmex.setup --mps
CPU-only (no GPU required, slower)
mix voxcpmex.setup --cpu
With virtual environment
mix voxcpmex.setup --cuda --venv .venv
## Quick Start
Start a model server (CUDA GPU)
= VoxCPMEx.start_link(device: "cuda")
Wait for model to load (30-60s on first run, downloads ~8GB)
:ok = VoxCPMEx.await_ready(pid)
Generate speech
= VoxCPMEx.generate(pid, "Hello, world from VoxCPM2!")
Save to file
:ok = VoxCPMEx.save(audio, "output.wav")
## Voice Design 🎨
Create a voice from a **text description** — no reference audio needed. Put the description in parentheses at the start of your text:
= VoxCPMEx.generate(pid, "(A young woman, gentle and sweet voice, warm tone) Hello, welcome to VoxCPM2!" )
Works with any descriptive language:
"(A deep male voice, authoritative and confident)" "(An elderly person, wise and slow-paced)" "(A cheerful child, energetic and bright)" "(A calm narrator, suitable for audiobooks)" "(A robot voice, mechanical and precise)" "(温柔甜美的少女声音)" # Chinese descriptions work too!
## Voice Cloning 🎛️
### Basic Cloning (reference-only)
= VoxCPMEx.generate(pid, "This is a cloned voice.", audio_prompt: "path/to/reference.wav" )
### Cloning with Style Control
= VoxCPMEx.generate(pid, "(slightly faster, cheerful tone) This clone has style guidance.", audio_prompt: "speaker.wav", cfg_value: 2.0, inference_timesteps: 10 )
### Ultimate Cloning (maximum fidelity)
= VoxCPMEx.generate(pid, "This is ultimate cloning.", prompt_wav_path: "speaker.wav", prompt_text: "The exact transcript of the reference audio.", audio_prompt: "speaker.wav" )
## Multilingual Support 🌍
VoxCPM2 is **tokenizer-free** — just feed text in any supported language, no language tag needed:
Chinese
= VoxCPMEx.generate(pid, "你好,今天天气真不错")
Japanese
= VoxCPMEx.generate(pid, "こんにちは、今日はいい天気ですね")
Korean
= VoxCPMEx.generate(pid, "안녕하세요, 오늘 날씨가 참 좋네요")
French
= VoxCPMEx.generate(pid, "Bonjour, le temps est magnifique aujourd'hui")
Arabic
= VoxCPMEx.generate(pid, "مرحبا، الطقس جميل اليوم")
## Quality Tuning
| Parameter | Range | Effect |
|-----------|-------|--------|
| `cfg_value` | 1.0–3.0 | Higher = stricter conditioning, less variation. Default: `2.0` |
| `inference_timesteps` | 4–30 | More steps = better detail, slower. Default: `10` |
High quality (more steps, slower)
= VoxCPMEx.generate(pid, "Quality matters.", inference_timesteps: 30, cfg_value: 3.0 )
Fast mode (fewer steps)
= VoxCPMEx.generate(pid, "Speed matters.", inference_timesteps: 4 )
## Streaming ⚡ (v2 — true chunk-by-chunk)
Start async streaming — returns immediately
= VoxCPMEx.generate_streaming_async(pid, "Long text for streaming...", inference_timesteps: 10, cfg_value: 2.0 )
Poll for chunks as they're generated
Stream.unfold(ref, fn ref -> case VoxCPMEx.next_chunk(pid, ref) do
{:ok, chunk} -> {chunk, ref} # got a chunk
:eos -> nil # stream done
{:error, reason} -> nil # errorend end) |> Enum.to_list()
Or collect all at once
= VoxCPMEx.generate_streaming_async(pid, "Long text...") {:ok, audio} = VoxCPMEx.collect_stream(pid, ref) :ok = VoxCPMEx.save(audio, "streaming.wav")
## Named Servers
Start with a name for easy access
= VoxCPMEx.start_link(device: "cuda", name: MyApp.TTS)
Use anywhere in your app
= VoxCPMEx.generate(MyApp.TTS, "Hello!")
## LoRA Fine-Tuning 🎓
Load fine-tuned weights
= VoxCPMEx.load_lora(pid, "path/to/lora_weights.ckpt")
Generate with adapted voice
= VoxCPMEx.generate(pid, "This uses my fine-tuned voice.")
Disable LoRA temporarily
:ok = VoxCPMEx.unload_lora(pid)
## Configuration
| Option | Description | Default |
|--------|-------------|---------|
| `:model` | HuggingFace model ID | `"openbmb/VoxCPM2"` |
| `:device` | Compute device (`"cuda"`, `"cpu"`, `"mps"`) | `"cuda"` |
| `:load_denoiser` | Load audio denoiser for reference cleanup | `false` |
| `:optimize` | Enable `torch.compile` | `true` |
| `:name` | GenServer name | `nil` |
## Generation Options
| Option | Description | Default |
|--------|-------------|---------|
| `:audio_prompt` | Reference audio for voice cloning | `nil` |
| `:prompt_wav_path` | Prompt audio for continuation cloning | `nil` |
| `:prompt_text` | Transcript of prompt audio | `nil` |
| `:cfg_value` | Guidance scale (1.0–3.0) | `2.0` |
| `:inference_timesteps` | Diffusion steps (4–30) | `10` |
| `:min_len` | Minimum audio length (tokens) | `2` |
| `:max_len` | Maximum token length | `4096` |
| `:normalize` | Text normalization | `false` |
| `:denoise` | Denoise reference audio | `false` |
## Supported Languages
Arabic, Burmese, Chinese (Mandarin + 四川话, 粤语, 吴语, 东北话, 河南话, 陕西话, 山东话, 天津话, 闽南话), Danish, Dutch, English, Finnish, French, German, Greek, Hebrew, Hindi, Indonesian, Italian, Japanese, Khmer, Korean, Lao, Malay, Norwegian, Polish, Portuguese, Russian, Spanish, Swahili, Swedish, Tagalog, Thai, Turkish, Vietnamese
## Hardware
| Device | VRAM | RTF (Speed) |
|--------|------|-------------|
| RTX 4090 | ~8 GB | 0.30 (standard) / 0.13 (Nano-VLLM) |
| RTX 3090 | ~8 GB | ~0.5 estimated |
| Apple M2 Max | Unified | Supported via MPS |
| CPU | N/A | Functional, much slower |
## License
Apache-2.0 — free for commercial use. VoxCPM2 model weights are also Apache-2.0.
⚠️ **Ethics:** Strictly forbidden to use for impersonation, fraud, or disinformation. AI-generated content should be clearly labeled.
## Links
- [VoxCPM2 on HuggingFace](https://huggingface.co/openbmb/VoxCPM2)
- [VoxCPM GitHub](https://github.com/OpenBMB/VoxCPM)
- [VoxCPM Docs](https://voxcpm.readthedocs.io)
- [Live Demo](https://huggingface.co/spaces/OpenBMB/VoxCPM-Demo)
- [Audio Samples](https://openbmb.github.io/voxcpm2-demopage)
- [Nano-VLLM Acceleration](https://github.com/a710128/nanovllm-voxcpm)
## Inspired By
This project follows the architecture pioneered by [chatterbex](https://github.com/holsee/chatterbex), an Elixir wrapper for Chatterbox TTS.