Optionalfields: ChatOpenAIFieldsOptional__Whether to include the raw OpenAI response in the output message's "additional_kwargs" field. Currently in experimental beta.
OptionalapiAPI key to use when making requests to OpenAI. Defaults to the value of
OPENAI_API_KEY environment variable.
OptionalaudioParameters for audio output. Required when audio output is requested with
modalities: ["audio"].
Learn more.
ProtectedcompletionsProtecteddefaultProtected OptionalfieldsOptionalfrequencyPenalizes repeated tokens according to frequency
OptionallogitDictionary used to adjust the probability of specific tokens being generated
OptionallogprobsWhether to return log probabilities of the output tokens or not. If true, returns the log probabilities of each output token returned in the content of message.
OptionalmaxMaximum number of tokens to generate in the completion. -1 returns as many tokens as possible given the prompt and the model's maximum context size.
OptionalmodalitiesOutput types that you would like the model to generate for this request. Most models are capable of generating text, which is the default:
["text"]
The gpt-4o-audio-preview model can also be used to
generate audio. To request that
this model generate both text and audio responses, you can use:
["text", "audio"]
Model name to use
OptionalmodelHolds any additional parameters that are valid to pass to openai.createCompletion that are not explicitly specified on this class.
OptionalnNumber of completions to generate for each prompt
OptionalorganizationOptionalpresencePenalizes repeated tokens
Used by OpenAI to cache responses for similar requests to optimize your cache hit rates. Learn more.
OptionalreasoningOptions for reasoning models.
Note that some options, like reasoning summaries, are only available when using the responses API. This option is ignored when not using a reasoning model.
ProtectedresponsesOptionalservice_Service tier to use for this request. Can be "auto", "default", or "flex" or "priority". Specifies the service tier for prioritization and latency optimization.
OptionalstopList of stop words to use when generating
Alias for stopSequences
OptionalstopList of stop words to use when generating
Whether to stream the results or not. Enabling disables tokenUsage reporting
Whether or not to include token usage data in streamed chunks.
OptionalsupportsWhether the model supports the strict argument when passing in tools.
If undefined the strict argument will not be passed to OpenAI.
OptionaltemperatureSampling temperature to use
OptionaltimeoutTimeout to use when making requests to OpenAI.
OptionaltopAn integer between 0 and 5 specifying the number of most likely tokens to return at each token position, each with an associated log probability. logprobs must be set to true if this parameter is used.
OptionaltopTotal probability mass of tokens to consider at each step
OptionaluserUnique string identifier representing your end-user, which can help OpenAI to monitor and detect abuse.
Whether to use the responses API for all requests. If false the responses API will be used
only when required in order to fulfill the request.
OptionalverbosityThe verbosity of the model's response.
OptionalzdrMust be set to true in tenancies with Zero Data Retention. Setting to true will disable
output storage in the Responses API, but this DOES NOT enable Zero Data Retention in your
OpenAI organization or project. This must be configured directly with OpenAI.
See: https://platform.openai.com/docs/guides/your-data https://platform.openai.com/docs/api-reference/responses/create#responses-create-store
Return profiling information for the model.
Provides information about the model's capabilities and constraints, including token limits, multimodal support, and advanced features like tool calling and structured output.
An object describing the model's capabilities and constraints
Protected_OptionaladditionalOptions: unknownProtected_Protected_Optionalkwargs: Partial<CallOptions>Optionaloptions: CallOptionsOptionaloptions: CallOptionsOptionalconfig: anyOptionalconfig: anyOptionalconfig: anyStaticlc_
OpenAI chat model integration.
To use with Azure, import the
AzureChatOpenAIclass.Setup: Install
@langchain/openaiand set an environment variable namedOPENAI_API_KEY.Constructor args
Runtime args
Runtime args can be passed as the second argument to any of the base runnable methods
.invoke..stream,.batch, etc. They can also be passed via.withConfig, or the second arg in.bindTools, like shown in the examples below:Examples
Instantiate
Invoking
Streaming Chunks
Aggregate Streamed Chunks
Bind tools
Structured Output
JSON Object Response Format
Multimodal
Usage Metadata
Logprobs
Response Metadata
JSON Schema Structured Output
Audio Outputs
Audio Outputs