VLLMOpenAI()Cohere async client.
Get the namespace of the langchain object.
Build extra kwargs from additional params that were passed in.
Validate environment.
Get the sub prompts for llm call.
Create the LLMResult from the choices and prompts.
Get the tokens present in the text with tiktoken package.
Calculate the maximum number of tokens possible to generate for a model.
Calculate the maximum number of tokens possible to generate for a prompt.
vLLM OpenAI-compatible API client