Inference API

Voice

View as Markdown

Create client secret

/v1/realtime/client_secrets

Create an ephemeral client secret for authenticating browser-side Realtime API connections.

Request Body

Response Body

value

string

The ephemeral token value. Use as a Bearer token in the WebSocket Authorization header, or in the sec-websocket-protocol header with prefix xai-client-secret..

expires_at

integer

Unix timestamp (seconds) when this client secret expires.


Realtime

WSSwss://api.x.ai/v1/realtime

Real-time voice conversations with Grok models via WebSocket. The connection begins with an HTTP GET that is upgraded to WebSocket (status 101). Once connected, the client and server exchange JSON messages to configure the session, stream audio, and receive responses.

Handshake

URL

wss://api.x.ai/v1/realtime

Method

GET

Status

101 Switching Protocols

Headers

Authorization

string

required

Bearer token for authentication. Use your xAI API key (server-side only) or an ephemeral client secret from the Create client secret endpoint.

Bearer $XAI_API_KEY

Sec-WebSocket-Protocol

string

Alternative authentication for browser clients. Pass the ephemeral token with prefix xai-client-secret.. When provided, the Authorization header is not required.

xai-client-secret.<EPHEMERAL_TOKEN>

Server → Client


Text to speech - REST

/v1/tts

Convert text into speech audio.

Request Body

text

string

required

The text to convert to speech. Maximum 15,000 characters. Supports inline speech tags for expressive output: [pause], [long-pause], [hum-tune], [laugh], [chuckle], [giggle], [cry], [tsk], [tongue-click], [lip-smack], [breath], [inhale], [exhale], [sigh]. Also supports wrapping tags for style control: <soft>, <whisper>, <loud>, <build-intensity>, <decrease-intensity>, <higher-pitch>, <lower-pitch>, <slow>, <fast>, <sing-song>, <singing>, <laugh-speak>, <emphasis>.

language

string

required

BCP-47 language code (e.g. en, zh, pt-BR) or auto for automatic language detection. Case-insensitive. Supported values: auto, en, ar-EG, ar-SA, ar-AE, bn, zh, fr, de, hi, id, it, ja, ko, pt-BR, pt-PT, ru, es-MX, es-ES, tr, vi. Additional languages may work with varying accuracy.


Streaming text to speech

WSSwss://api.x.ai/v1/tts

Bidirectional streaming text-to-speech via WebSocket. Send text incrementally and receive audio chunks in real time. Shares the /v1/tts path with the batch POST endpoint — a GET with Upgrade: websocket activates streaming mode. Configuration is done via query parameters at connection time. Supports multi-utterance: after audio.done, send another stream of text.delta messages on the same connection.

Handshake

URL

wss://api.x.ai/v1/tts

Method

GET

Status

101 Switching Protocols

Headers

Authorization

string

required

Bearer token for authentication. Use your xAI API key.

Bearer $XAI_API_KEY

Query Parameters

voice

string

optional

Voice identifier. Case-insensitive.

eveararexsalleo

language

string

required

BCP-47 language code (e.g. `en`, `zh`, `pt-BR`) or `auto` for automatic language detection. Case-insensitive.

autoenar-EGar-SAar-AEbnzhfrdehiiditjakopt-BRpt-PTrues-MXes-EStrvi

codec

string

optional

Audio codec for the output.

mp3wavpcmmulawalaw

sample_rate

integer

optional

Sample rate in Hz.

80001600022050240004410048000

bit_rate

integer

optional

Bit rate in bps. Only applies when `codec` is `mp3`.

320006400096000128000192000

Handshake

URLwss://api.x.ai/v1/tts
MethodGET
Status101 Switching Protocols

Text to speech - List voices

/v1/tts/voices

List all available TTS voices.

Response Body

voices

array

List of available voices.


Text to speech - Get voice

/v1/tts/voices/{voice_id}

Get details for a specific voice.

Path parameters

voice_id

string

required

The unique identifier of the voice (e.g. `eve`, `ara`). Case-insensitive.

Response Body

voice_id

string

Unique identifier for the voice (lowercase). Pass this value as voice_id in TTS requests or as the voice parameter in Realtime API session configuration.

name

string

Human-readable display name for the voice.


Speech to text - REST

/v1/stt

Transcribe an audio file to text.

Request Body

Response Body

text

string

Full transcript text. For multichannel requests, this is a merged transcript across all channels (words interleaved by timestamp).

language

string

Detected language code (ISO 639-1, e.g. en). Currently empty — language detection is not yet enabled.

duration

number

Audio duration in seconds (rounded to 2 decimal places).

words

array

Word-level segments with timestamps. Omitted when empty.

channels

array

Per-channel transcripts. Only present when multichannel=true. Omitted for single-channel audio.


Speech to text - Streaming

WSSwss://api.x.ai/v1/stt

Real-time streaming speech-to-text via WebSocket. The client streams raw audio as binary WebSocket frames, and the server returns JSON transcript events as the audio is processed.
Configuration is done via query parameters on the WebSocket upgrade URL. Audio is sent as raw binary frames — no base64 encoding needed.
The server uses VAD (Voice Activity Detection) to detect when the speaker stops talking and emits utterance-final events. For long continuous speech, the server automatically splits into ~3-second chunks.
After sending audio.done, the server returns a transcript.done event with any remaining transcript not already covered by speech_final events, plus the total audio duration. If all audio was covered by speech_final events, text and words will be empty. The client can then start a new turn without reconnecting.

Handshake

URL

wss://api.x.ai/v1/stt

Method

GET

Status

101 Switching Protocols

Headers

Authorization

string

required

Bearer token authentication. Format: Bearer <your xAI API key>.

Bearer $XAI_API_KEY

Query Parameters

sample_rate

integer

optional

Audio sample rate in Hz. Supported values: `8000`, `16000`, `22050`, `24000`, `44100`, `48000`.

encoding

string

optional

Audio encoding format. `pcm` — signed 16-bit little-endian (2 bytes/sample). `mulaw` — G.711 µ-law (1 byte/sample). `alaw` — G.711 A-law (1 byte/sample).

interim_results

boolean

optional

When `true`, the server emits partial transcript events (`is_final=false`) approximately every 500 ms while audio is being processed. When `false` (default), only finalized results are sent.

endpointing

integer

optional

Silence duration in milliseconds before the server fires a `speech_final=true` event, indicating the speaker stopped talking. Range: 0–5000. Set to `0` for no delay (fire on any VAD silence boundary). Default: 10ms.

language

string

optional

Language code (e.g. `en`, `fr`, `de`, `ja`). When set, enables Inverse Text Normalization — spoken-form numbers, currencies, and units are converted to their written form.

multichannel

boolean

optional

When `true`, enables per-channel transcription for interleaved multichannel audio. Requires `channels` to be set to ≥ 2.

channels

integer

optional

Number of interleaved audio channels. Required when `multichannel=true`. Min: 2, Max: 8.

diarize

boolean

optional

When `true`, enables speaker diarization. Words in `transcript.partial` and `transcript.done` events include a `speaker` field (integer) identifying the detected speaker.

Server → Client


Did you find this page helpful?

Last updated: April 10, 2026