Create a new response using the Response API. This endpoint provides an OpenAI Responses-compatible interface for chat completions with video collections.
The response can be processed synchronously (default) or asynchronously using the background parameter.
Bearer authentication header of the form Bearer <token>, where <token> is your auth token.
Response creation parameters
The model to use for generating the response.
nimbus-001: Fast general question answering model. Requires knowledge_base. Default temperature: 0.7.nimbus-002-preview: Light reasoning model capable of multi-step reasoning, cross-video synthesis, and structured entity data. Requires knowledge_base. Supports entity_collections in the knowledge base. Default temperature: 1.1"nimbus-001"
The input for the response. Can be a simple string or an array of messages.
System instructions to guide the model's behavior (maps to developer/system message)
Sampling temperature for the model. Defaults to 0.7 for nimbus-001 and 1 for nimbus-002-preview.
0 <= x <= 2Knowledge base configuration. Required for nimbus-001 and nimbus-002-preview.
Additional data to include in the response annotations
cloudglue_citations.media_descriptions Tool definitions for function calling.
Maximum number of output tokens.
1 <= x <= 128000Set to true to process the response in the background. When true, the response is returned immediately with status 'in_progress'.
Stream response via SSE. Mutually exclusive with background.
Response created successfully
Unique identifier for the response
Object type identifier
response Current status of the response
in_progress, completed, failed, cancelled Unix timestamp of when the response was created
The model used for the response
The system instructions used
The generated output messages
Token usage statistics for the response
Error details when the response status is 'failed'