# google.generativeai.protos.GenerateAnswerRequest Request to generate a grounded answer from the ``Model``. This message has `oneof`_ fields (mutually exclusive fields). For each oneof, at most one member field can be set at the same time. Setting any member of the oneof automatically clears all other members.
`inline_passages` `google.ai.generativelanguage.GroundingPassages` Passages provided inline with the request. This field is a member of `oneof`_ ``grounding_source``.
`semantic_retriever` `google.ai.generativelanguage.SemanticRetrieverConfig` Content retrieved from resources created via the Semantic Retriever API. This field is a member of `oneof`_ ``grounding_source``.
`model` `str` Required. The name of the ``Model`` to use for generating the grounded response. Format: ``model=models/{model}``.
`contents` `MutableSequence[google.ai.generativelanguage.Content]` Required. The content of the current conversation with the ``Model``. For single-turn queries, this is a single question to answer. For multi-turn queries, this is a repeated field that contains conversation history and the last ``Content`` in the list containing the question. Note: ``GenerateAnswer`` only supports queries in English.
`answer_style` `google.ai.generativelanguage.GenerateAnswerRequest.AnswerStyle` Required. Style in which answers should be returned.
`safety_settings` `MutableSequence[google.ai.generativelanguage.SafetySetting]` Optional. A list of unique ``SafetySetting`` instances for blocking unsafe content. This will be enforced on the GenerateAnswerRequest.contents and ``GenerateAnswerResponse.candidate``. There should not be more than one setting for each ``SafetyCategory`` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each ``SafetyCategory`` specified in the safety_settings. If there is no ``SafetySetting`` for a given ``SafetyCategory`` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT are supported. Refer to the `guide `__ for detailed information on available safety settings. Also refer to the `Safety guidance `__ to learn how to incorporate safety considerations in your AI applications.
`temperature` `float` Optional. Controls the randomness of the output. Values can range from [0.0,1.0], inclusive. A value closer to 1.0 will produce responses that are more varied and creative, while a value closer to 0.0 will typically result in more straightforward responses from the model. A low temperature (~0.2) is usually recommended for Attributed-Question-Answering use cases.
## Child Classes [`class AnswerStyle`](../../../google/generativeai/protos/GenerateAnswerRequest/AnswerStyle.md)