Generate audio from text input while managing multiple independent audio generation streams over a single WebSocket connection.
The independent audio streams each correspond to a context, identified by contextId, that maintains its own state. To use the API:
maxBufferDelayMs and bufferCharThreshold in the context configurations).auto_mode which would automatically balance latency and quality of the generations.contextId so you can match the audio to the request.