Optionalfrequency_Optionallogit_JSON object that maps tokens to an associated bias value from -100 to 100.
Optionalmax_The maximum number of tokens to generate in the chat completion.
A list of messages comprising the conversation so far.
The content of the message.
The role of the sender (e.g., 'user' or 'assistant').
user - userassistant - assistantOptionalmodelID of the model to use. See the model endpoint compatibility table for details.
Optionalpresence_Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far.
The ID of the project to use.
OptionalrepositoriesOptions for Retrieval Augmented Generation (RAG). Will override launched model settings
Optionalids?: number[]The IDs of the repositories to use.
Optionallimit?: numberOptionalsimilarity_Optionalresponse_An object specifying the format that the model must output.
OptionalseedThis feature is in Beta. If specified, our system will make a best effort to sample deterministically.
Optionalsession_The ID of the session to use. It helps to track the chat history.
OptionalstopUp to 4 sequences where the API will stop generating further tokens.
If set, partial message deltas will be sent, like in ChatGPT.
Optionalsystem_The system prompt to use.
OptionaltemperatureWhat sampling temperature to use, between 0 and 2.
OptionaltoolsA list of tools the model may call. Currently, only functions are supported as a tool.
Optionaltop_An alternative to sampling with temperature, called nucleus sampling.
OptionaluserA unique identifier representing your end-user.
Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency.