Optional
fields: AzureChatOpenAIFieldsOptional
__Whether to include the raw OpenAI response in the output message's "additional_kwargs" field. Currently in experimental beta.
Optional
apiAPI key to use when making requests to OpenAI. Defaults to the value of
OPENAI_API_KEY
environment variable.
Optional
audioParameters for audio output. Required when audio output is requested with
modalities: ["audio"]
.
Learn more.
Optional
azureA function that returns an access token for Microsoft Entra (formerly known as Azure Active Directory), which will be invoked on every request.
Optional
azureAzure OpenAI API deployment name to use for completions when making requests to Azure OpenAI. This is the name of the deployment you created in the Azure portal. e.g. "my-openai-deployment" this will be used in the endpoint URL: https://{InstanceName}.openai.azure.com/openai/deployments/my-openai-deployment/
Optional
azureAzure OpenAI API instance name to use when making requests to Azure OpenAI. this is the name of the instance you created in the Azure portal. e.g. "my-openai-instance" this will be used in the endpoint URL: https://my-openai-instance.openai.azure.com/openai/deployments/{DeploymentName}/
Optional
azureAPI key to use when making requests to Azure OpenAI.
Optional
azureAPI version to use when making requests to Azure OpenAI.
Optional
azureCustom base url for Azure OpenAI API. This is useful in case you have a deployment in another region. e.g. setting this value to "https://westeurope.api.cognitive.microsoft.com/openai/deployments" will be result in the endpoint URL: https://westeurope.api.cognitive.microsoft.com/openai/deployments/{DeploymentName}/
Optional
azureCustom endpoint for Azure OpenAI API. This is useful in case you have a deployment in another region. e.g. setting this value to "https://westeurope.api.cognitive.microsoft.com/" will be result in the endpoint URL: https://westeurope.api.cognitive.microsoft.com/openai/deployments/{DeploymentName}/
Protected
completionsProtected
defaultProtected
Optional
fieldsOptional
frequencyPenalizes repeated tokens according to frequency
Optional
logitDictionary used to adjust the probability of specific tokens being generated
Optional
logprobsWhether to return log probabilities of the output tokens or not. If true, returns the log probabilities of each output token returned in the content of message.
Optional
maxMaximum number of tokens to generate in the completion. -1 returns as many tokens as possible given the prompt and the model's maximum context size.
Optional
modalitiesOutput types that you would like the model to generate for this request. Most models are capable of generating text, which is the default:
["text"]
The gpt-4o-audio-preview
model can also be used to
generate audio. To request that
this model generate both text and audio responses, you can use:
["text", "audio"]
Model name to use
Optional
modelHolds any additional parameters that are valid to pass to openai.createCompletion
that are not explicitly specified on this class.
Optional
nNumber of completions to generate for each prompt
Optional
organizationOptional
presencePenalizes repeated tokens
Used by OpenAI to cache responses for similar requests to optimize your cache hit rates. Learn more.
Optional
reasoningOptions for reasoning models.
Note that some options, like reasoning summaries, are only available when using the responses API. This option is ignored when not using a reasoning model.
Protected
responsesOptional
service_Service tier to use for this request. Can be "auto", "default", or "flex" or "priority". Specifies the service tier for prioritization and latency optimization.
Optional
stopList of stop words to use when generating
Alias for stopSequences
Optional
stopList of stop words to use when generating
Whether to stream the results or not. Enabling disables tokenUsage reporting
Whether or not to include token usage data in streamed chunks.
Optional
supportsWhether the model supports the strict
argument when passing in tools.
If undefined
the strict
argument will not be passed to OpenAI.
Optional
temperatureSampling temperature to use
Optional
timeoutTimeout to use when making requests to OpenAI.
Optional
topAn integer between 0 and 5 specifying the number of most likely tokens to return at each token position, each with an associated log probability. logprobs must be set to true if this parameter is used.
Optional
topTotal probability mass of tokens to consider at each step
Optional
userUnique string identifier representing your end-user, which can help OpenAI to monitor and detect abuse.
Whether to use the responses API for all requests. If false
the responses API will be used
only when required in order to fulfill the request.
Optional
verbosityThe verbosity of the model's response.
Optional
zdrMust be set to true
in tenancies with Zero Data Retention. Setting to true
will disable
output storage in the Responses API, but this DOES NOT enable Zero Data Retention in your
OpenAI organization or project. This must be configured directly with OpenAI.
See: https://platform.openai.com/docs/guides/your-data https://platform.openai.com/docs/api-reference/responses/create#responses-create-store
Protected
_Optional
additionalOptions: unknownProtected
_Protected
_Optional
kwargs: Partial<CallOptions>Optional
options: CallOptionsOptional
options: CallOptionsOptional
config: anyOptional
config: anyOptional
config: anyStatic
lc_
Azure OpenAI chat model integration.
Setup: Install
@langchain/openai
and set the following environment variables:Constructor args
Runtime args
Runtime args can be passed as the second argument to any of the base runnable methods
.invoke
..stream
,.batch
, etc. They can also be passed via.withConfig
, or the second arg in.bindTools
, like shown in the examples below:Examples
Instantiate
Invoking
Streaming Chunks
Aggregate Streamed Chunks
Bind tools
Structured Output
JSON Object Response Format
Multimodal
Usage Metadata
Logprobs
Response Metadata