Parameters for the GoogleVertexAIMultimodalEmbeddings class, extending both EmbeddingsParams and GoogleVertexAIConnectionParams.
interface GoogleVertexAIMultimodalEmbeddingsParamsEmbeddingsParamsGoogleVertexAIBaseLLMInput<GoogleAuthOptions>The version of the API functions. Part of the path.
If you are planning to connect to a model that lives under a custom endpoint provide the "customModelURL" which will override the automatic URL building
This is necessary in cases when you want to point to a fine-tuned model or a model that has been hidden under VertexAI Endpoints.
In those cases, specifying the GoogleVertexAIModelParams.model param
will not be necessary and will be ignored.
Hostname for the API call
Region where the LLM is stored
The maximum number of concurrent calls that can be made.
Defaults to Infinity, which means no limit.
Maximum number of tokens to generate in the completion.
The maximum number of retries that can be made for a single call, with an exponential backoff between each attempt. Defaults to 6.
Model name to use. Available options are: qwen-turbo, qwen-plus, qwen-max, or Other compatible models.
Custom handler to handle failed attempts. Takes the originally thrown error object as input, and should itself throw an error if the input error is not retryable.
Amount of randomness injected into the response. Ranges from 0 to 1 (0 is not included). Use temp closer to 0 for analytical / multiple choice, and temp closer to 1 for creative and generative tasks. Defaults to 0.95.
Total probability mass of tokens to consider at each step. Range from 0 to 1.0. Defaults to 0.8.
Whether to print out response text.