deploy
parameter to true
.
Get the prompt_id
from the prompts endpoint.
The list of messages for the prompt version. If you want to add a variable, you can use the following format:
{{variable_name}}
.Example
Example
Speciy the model you want to use in this version.
Description of the prompt version
Whether the prompt version should be streamed or not. Default is
false
.The temperature of the model.
The maximum number of tokens to generate.
The nucleus sampling probability.
Specify how much to penalize new tokens based on their existing frequency in the text so far. Decreases the model’s likelihood of repeating the same line verbatim
Specify how much to penalize new tokens based on whether they appear in the text so far. Increases the model’s likelihood of talking about new topics
The list of variables for the prompt version. You can use these variables in the messages.
Example
Example
The list of fallback models for the prompt version. Check out fallback models for more information.
Example
Example
The list of models to load balance the prompt version. Check out load balancing for more information.
Example
Example
The list of tools to use for the prompt version. Check out tools for more information.
Example
Example
Whether to deploy this version as the live version Default is
false
.