GLM-TTS
GLM-TTS
Provider: Zhipu AI
Category: Audio generation
Endpoint: POST /v1/audio/speech
Context window: —
Served from: EmpirioLabs (Native Inference)
LLM-based text-to-speech with zero-shot voice cloning from 3-10s of audio and emotion-expressive, controllable output via multi-reward RL.
At a glance
Pricing
Example request
Parameters
Every parameter this model accepts is documented in the live machine-readable schema returned by GET /v1/models/glm-tts. Common controls include temperature, top_p, max_tokens, and the universal disable_formatting passthrough flag (also accepted as raw=true).
