langchain.js
    Preparing search index...

    Interface GoogleGenerativeAIChatInput

    An interface defining the input to the ChatGoogleGenerativeAI class.

    interface GoogleGenerativeAIChatInput {
        apiKey?: string;
        apiVersion?: string;
        baseUrl?: string;
        convertSystemMessageToHumanContent?: boolean;
        json?: boolean;
        maxOutputTokens?: number;
        model: string;
        safetySettings?: SafetySetting[];
        stopSequences?: string[];
        streaming?: boolean;
        streamUsage?: boolean;
        temperature?: number;
        topK?: number;
        topP?: number;
    }

    Hierarchy (View Summary)

    Implemented by

    Index

    Properties

    apiKey?: string

    Google API key to use

    apiVersion?: string

    Google API version to use

    baseUrl?: string

    Google API base URL to use

    convertSystemMessageToHumanContent?: boolean

    Whether or not model supports system instructions. The following models support system instructions:

    • All Gemini 1.5 Pro model versions
    • All Gemini 1.5 Flash model versions
    • Gemini 1.0 Pro version gemini-1.0-pro-002
    json?: boolean

    Whether or not to force the model to respond with JSON. Available for gemini-1.5 models and later.

    false
    
    maxOutputTokens?: number

    Maximum number of tokens to generate in the completion.

    model: string

    Model Name to use

    Note: The format must follow the pattern - {model}

    safetySettings?: SafetySetting[]

    A list of unique SafetySetting instances for blocking unsafe content. The API will block any prompts and responses that fail to meet the thresholds set by these settings. If there is no SafetySetting for a given SafetyCategory provided in the list, the API will use the default safety setting for that category.

    stopSequences?: string[]

    The set of character sequences (up to 5) that will stop output generation. If specified, the API will stop at the first appearance of a stop sequence.

    Note: The stop sequence will not be included as part of the response. Note: stopSequences is only supported for Gemini models

    streaming?: boolean

    Whether to stream the results or not

    streamUsage?: boolean

    Whether or not to include usage data, like token counts in the streamed response chunks.

    true
    
    temperature?: number

    Controls the randomness of the output.

    Values can range from [0.0,2.0], inclusive. A value closer to 2.0 will produce responses that are more varied and creative, while a value closer to 0.0 will typically result in less surprising responses from the model.

    Note: The default value varies by model

    topK?: number

    Top-k changes how the model selects tokens for output.

    A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature).

    Note: The default value varies by model

    topP?: number

    Top-p changes how the model selects tokens for output.

    Tokens are selected from most probable to least until the sum of their probabilities equals the top-p value.

    For example, if tokens A, B, and C have a probability of .3, .2, and .1 and the top-p value is .5, then the model will select either A or B as the next token (using temperature).

    Note: The default value varies by model