GigaChat large language models API.
Base class of Friendli.
Friendli LLM.
Parameters for the Javelin AI Gateway LLM.
Javelin AI Gateway LLMs.
Writer large language models.
Parameters for AI21 penalty data.
AI21 large language models.
Fake LLM for testing purposes.
Fake streaming list LLM for testing purposes.
Banana large language models.
HazyResearch's Manifest library.
Wrapper around You.com's conversational Smart and Research APIs.
Weight only quantized model.
The device to use for inference, cuda or cpu
Configuration for the reader to be deployed in Titan Takeoff API.
Titan Takeoff API LLMs.
Yi large language models.
Parse the byte stream input.
Handler class to transform input from LLM to a
Content handler for LLM class.
Use your Predibase models with Langchain.
Aphrodite language model.
MLX Pipeline API.
Parameters for the MLflow AI Gateway LLM.
MLflow AI Gateway LLMs.
Nebula Service models.
OCI authentication types as enumerator.
Base class for OCI GenAI models
OCI large language models.
AzureML Managed Endpoint client.
Azure ML endpoints API types. Use dedicated for models deployed in hosted
Transform request and response of AzureML endpoint to match with
Content handler for GPT2
Deprecated: Kept for backwards compatibility
Content handler for LLMs from the HuggingFace catalog.
Content handler for the Dolly-v2-12b model
Content formatter for models that use the OpenAI like API scheme.
Deprecated: Kept for backwards compatibility
Azure ML Online Endpoint models.
Azure ML Online Endpoint models.
Common parameters for Minimax large language models.
Minimax large language models.
Common parameters for Moonshot LLMs.
Moonshot large language models.
Base OpenAI large language model class.
Langchain LLM class to help to access eass llm service.
Baichuan large language models.
iFlyTek Spark completion model integration.
EdenAI models.
SambaStudio large language models.
SambaNova Cloud large language models.
NLPCloud large language models.
MosaicML LLM service.
GooseAI large language models.
Replicate models.
MLflow LLM service.
LLM that uses OpaquePrompts to sanitize prompts.
llama.cpp model.
Anyscale large language models.
User input as the response.
Petals Bloom models.
Aviary backend.
Aviary hosted models.
ForefrontAI large language models.
Layerup Security LLM service.
Kobold API language model.
ExllamaV2 API.
Modal large language models.
NIBittensor LLMs
Common configuration for Solar LLMs.
Solar large language models.
Raised when the Ollama endpoint is not found.
ChatGLM3 LLM service.
Base class for VolcEngineMaas models.
volc engine maas hosts a plethora of models.
Neural Magic DeepSparse LLM interface.
HuggingFace Pipeline API to run on self-hosted remote hardware.
Adapter class to prepare the inputs from Langchain to a format
Base class for Bedrock models.
Baseten model
Model inference on self-hosted remote hardware.
Konko AI models.
CTranslate2 language model.
PipelineAI large language models.
Tongyi completion model integration.
Cloudflare Workers AI service.
Text generation models from WebUI.
Yandex large language models.
GPT4All language models.
Train result.
Gradient.ai LLM Endpoints.
Llamafile lets you distribute and run large language models with a
StochasticAI large language models.
OctoAI LLM Endpoints - OpenAI compatible.
RWKV language models.
LLM wrapper for the Outlines library.
CerebriumAI large language models.
OpenLM models.
Wrapper around the BigdlLLM model
Adapter to prepare the inputs from Langchain to a format
Amazon API Gateway to access LLM models hosted on AWS.
Xinference large-scale model inference service.
Baidu Qianfan completion model integration.
VLLM language model.
vLLM OpenAI-compatible API client
Beam API for gpt2 large language model.
DeepInfra models.
OpenAI's compatible API client for OpenLLM server
Raises when token expired.
Raises when encounter server error when making inference.
Base class for LLM deployed on OCI Data Science Model Deployment.
LLM deployed on OCI Data Science Model Deployment.
OCI Data Science Model Deployment TGI Endpoint.
VLLM deployed on OCI Data Science Model Deployment
ChatGLM LLM service.
Arcee's Domain Adapted Language Models (DALMs).
Yuan2.0 language models.
C Transformers LLM models.
Aleph Alpha large language models.
Clarifai large language models.
PromptLayer OpenAI large language models.
PromptLayer OpenAI large language models.
IpexLLM model.
Base class for Cohere models.
Cohere large language models.
Sagemaker Inference Endpoint models.
DEPRECATED: Use langchain_google_genai.GoogleGenerativeAI instead.
OpenAI large language models.
Azure-specific OpenAI large language models.
OpenAI Chat large language models.
HuggingFace text generation API.
Fireworks models.
Google Vertex AI large language models.
Vertex AI Model Garden large language models.
Ollama locally runs large language models.
Prediction Guard large language models.
HuggingFace Endpoint.
Bedrock models.
HuggingFaceHub models.
Databricks serving endpoint or a cluster driver proxy app for LLM.
HuggingFace Pipeline API.
IBM watsonx.ai large language models.
LLM models from Together.
Anthropic large language models.
Use tenacity to retry the completion call.
Use tenacity to retry the completion call.
Load LLM from Config Dict.
Load LLM from a file.
Use tenacity to retry the completion call.
Generate text from the model.
Use tenacity to retry the completion call.
Cut off the text as soon as any stop words occur.
Update token usage.
Use tenacity to retry the completion call.
Use tenacity to retry the async completion call.
Update token usage.
Create the LLMResult from the choices and prompts.
List available models
Get completions from Aviary models.
Conditionally apply a decorator.
Use tenacity to retry the completion call.
Use tenacity to retry the completion call.
Use tenacity to retry the completion call.
Use tenacity to retry the completion call.
Use tenacity to retry the completion call for streaming.
Default guardrail violation handler.
Remove trailing slash and /api from url if present.
Return True if the model name is a Codey model.
Return True if the model name is a Gemini model.
Use tenacity to retry the completion call.
Use tenacity to retry the completion call.
Get the notebook REPL context if running inside a Databricks notebook.
Get the default Databricks workspace hostname.
Get the default Databricks personal access token.
Check the response from the completion call.
Use tenacity to retry the completion call.
Use tenacity to retry the completion call.
Async version of stream_generate_with_retry.
Generate elements from an iterable,
Generate elements from an async iterable,
Use tenacity to retry the completion call.
Use tenacity to retry the async completion call.