Skip to content

Latest commit

 

History

History
142 lines (86 loc) · 2.28 KB

GenerateMessageRequest.md

File metadata and controls

142 lines (86 loc) · 2.28 KB

google.generativeai.protos.GenerateMessageRequest

View source on GitHub

Request to generate a message response from the model.

Attributes

model

str

Required. The name of the model to use.

Format: name=models/{model}.

prompt

google.ai.generativelanguage.MessagePrompt

Required. The structured textual input given to the model as a prompt. Given a prompt, the model will return what it predicts is the next message in the discussion.

temperature

float

Optional. Controls the randomness of the output.

Values can range over [0.0,1.0], inclusive. A value closer to 1.0 will produce responses that are more varied, while a value closer to 0.0 will typically result in less surprising responses from the model.

candidate_count

int

Optional. The number of generated response messages to return.

This value must be between [1, 8], inclusive. If unset, this will default to 1.

top_p

float

Optional. The maximum cumulative probability of tokens to consider when sampling.

The model uses combined Top-k and nucleus sampling.

Nucleus sampling considers the smallest set of tokens whose probability sum is at least top_p.

top_k

int

Optional. The maximum number of tokens to consider when sampling.

The model uses combined Top-k and nucleus sampling.

Top-k sampling considers the set of top_k most probable tokens.