Interface defining the parameters for the Perplexity chat model.
interface PerplexityChatInputWhen set to true, disables web search completely and the model will only use its training data to respond. This is useful when you want deterministic responses without external information.
Enables a classifier that decides if web search is needed based on your query.
Penalizes repeated tokens according to frequency
Filters search results to only include content last updated after this date.
Filters search results to only include content last updated before this date.
The maximum number of concurrent calls that can be made.
Defaults to Infinity, which means no limit.
The maximum number of retries that can be made for a single call, with an exponential backoff between each attempt. Defaults to 6.
Model name to use. Available options are: qwen-turbo, qwen-plus, qwen-max, or Other compatible models.
Custom handler to handle failed attempts. Takes the originally thrown error object as input, and should itself throw an error if the input error is not retryable.
Penalizes repeated tokens
A lower reasoning effort can result in faster responses, fewer tokens used, and shorter reasoning_content in the responses. Supported values are low, medium, and high.
Whether to return images
Determines whether or not a request to an online model should return related questions.
Filters search results to only include content published after this date.
Filters search results to only include content published before this date.
Search domain filter - limit the citations used by the online model to URLs from the specified domains.
Controls the search mode used for the request. When set to 'academic', results will prioritize scholarly sources.
Returns search results within the specified time interval - does not apply to images. Values include month, week, day, hour.
Whether to stream the results or not. Defaults to false.
Amount of randomness injected into the response. Ranges from 0 to 1 (0 is not included). Use temp closer to 0 for analytical / multiple choice, and temp closer to 1 for creative and generative tasks. Defaults to 0.95.
Timeout for this call in milliseconds.
Total probability mass of tokens to consider at each step. Range from 0 to 1.0. Defaults to 0.8.
Whether to print out response text.
Configuration for using web search in model responses.