Maximum number of tokens to return in the response.
List of strings that define sequences after which the model will stop generating.
Controls the randomness of the response. Choose a lower value for more predictable outputs and a higher value for more surprising outputs.
Percentage of most-likely candidates that the model considers for the next token.