interface ChatTogetherAICallOptionsOmit<ChatOpenAICallOptions, TogetherAIUnsupportedCallOptions>Parameters for audio output. Required when audio output is requested with
modalities: ["audio"].
Learn more.
Runtime values for attributes previously made configurable on this Runnable, or sub-Runnables.
Specify additional output data to include in the model response.
Describes the format of structured outputs. This should be provided if an output is considered to be structured
The maximum number of concurrent calls that can be made.
Defaults to Infinity, which means no limit.
Output types that you would like the model to generate for this request. Most models are capable of generating text, which is the default:
["text"]
The gpt-4o-audio-preview model can also be used to
generate audio. To request that
this model generate both text and audio responses, you can use:
["text", "audio"]
Additional options to pass to the underlying axios request.
The model may choose to call multiple functions in a single turn. You can set parallel_tool_calls to false which ensures only one tool is called at most. Learn more
Static predicted output content, such as the content of a text file that is being regenerated. Learn more.
The unique ID of the previous response to the model. Use this to create multi-turn conversations.
Used by OpenAI to cache responses for similar requests to optimize your cache hit rates. Learn more.
Used by OpenAI to set cache retention time
Options for reasoning models.
Note that some options, like reasoning summaries, are only available when using the responses API. This option is ignored when not using a reasoning model.
A lower reasoning effort can result in faster responses, fewer tokens used, and shorter reasoning_content in the responses. Supported values are low, medium, and high.
Maximum number of times a call can recurse. If not provided, defaults to 25.
An object specifying the format that the model must output.
Unique identifier for the tracer run for this call. If not provided, a new UUID will be generated.
Name for the tracer run for this call. Defaults to the name of the class.
Service tier to use for this request. Can be "auto", "default", or "flex" or "priority". Specifies the service tier for prioritization and latency optimization.
Abort signal for this call. If provided, the call will be aborted when the signal is aborted.
Stop tokens to use for this call. If not provided, the default stop tokens for the model will be used.
Additional options to pass to streamed completions. If provided, this takes precedence over "streamUsage" set at initialization time.
If true, model output is guaranteed to exactly match the JSON Schema
provided in the tool definition. If true, the input schema will also be
validated according to
https://platform.openai.com/docs/guides/structured-outputs/supported-schemas.
If false, input schema will not be validated and model output will not
be validated.
If undefined, strict argument will not be passed to the model.
Configuration options for a text response from the model. Can be plain text or structured JSON data.
Timeout for this call in milliseconds.
Specifies how the chat model should use tools.
The truncation strategy to use for the model response.
The verbosity of the model's response.