User input for a chat completion request.

model
enum<string>
required

A model name, for example 'sutra-light'.

Available options:
sutra-light,
sutra-pro,
sutra-online
messages
object[]
required

The LLM prompt.

max_tokens
number

The maximum number of tokens to generate before terminating. This number cannot exceed the context window for the selected model. The default value is 1024.

temperature
number

Controls the randomness of the response, a lower temperature gives lower randomness. Values are in the range [0,2] with a default value of 0.3.

stop
object

May be a string, null or an array of strings.

presence_penalty
number
frequency_penalty
number
top_p
number
extra_body
object