Responses
OpenResponses Types
Data models and types for the OpenResponses API specification.
For the full OpenResponses type definitions, see the openresponses-types package documentation.
any_llm.types.responses
ResponsesParams
Bases: BaseModel
Normalized parameters for responses API.
This model is used internally to pass structured parameters from the public API layer to provider implementations, avoiding very long function signatures while keeping type safety.
Source code in src/any_llm/types/responses.py
15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 | |
background = None
class-attribute
instance-attribute
Whether to run the request in the background and return immediately.
conversation = None
class-attribute
instance-attribute
The conversation to associate this response with (ID string or ConversationParam object).
frequency_penalty = None
class-attribute
instance-attribute
Penalizes new tokens based on their frequency in the text so far.
include = None
class-attribute
instance-attribute
Items to include in the response (e.g., 'reasoning.encrypted_content').
input
instance-attribute
The input payload accepted by provider's Responses API. For OpenAI-compatible providers, this is typically a list mixing text, images, and tool instructions, or a dict per OpenAI spec.
max_output_tokens = None
class-attribute
instance-attribute
Maximum number of tokens to generate
metadata = None
class-attribute
instance-attribute
Key-value pairs for custom metadata (up to 16 pairs).
model
instance-attribute
Model identifier (e.g., 'mistral-small-latest')
parallel_tool_calls = None
class-attribute
instance-attribute
Whether to allow parallel tool calls
presence_penalty = None
class-attribute
instance-attribute
Penalizes new tokens based on whether they appear in the text so far.
previous_response_id = None
class-attribute
instance-attribute
The ID of the response to use as the prior turn for this request.
prompt_cache_key = None
class-attribute
instance-attribute
A key to use when reading from or writing to the prompt cache.
prompt_cache_retention = None
class-attribute
instance-attribute
How long to retain a prompt cache entry created by this request.
reasoning = None
class-attribute
instance-attribute
Configuration options for reasoning models.
response_format = None
class-attribute
instance-attribute
Format specification for the response
safety_identifier = None
class-attribute
instance-attribute
A stable identifier used for safety monitoring and abuse detection.
service_tier = None
class-attribute
instance-attribute
The service tier to use for this request.
store = None
class-attribute
instance-attribute
Whether to store the response so it can be retrieved later.
stream = None
class-attribute
instance-attribute
Whether to stream the response
stream_options = None
class-attribute
instance-attribute
Additional options controlling streaming behavior
temperature = None
class-attribute
instance-attribute
Controls randomness in the response (0.0 to 2.0)
tool_choice = None
class-attribute
instance-attribute
Controls which tools the model can call
tools = None
class-attribute
instance-attribute
List of tools for tool calling. Should be converted to OpenAI tool format dicts
top_logprobs = None
class-attribute
instance-attribute
Number of top alternatives to return when logprobs are requested
top_p = None
class-attribute
instance-attribute
Controls diversity via nucleus sampling (0.0 to 1.0)
truncation = None
class-attribute
instance-attribute
Controls how the service truncates the input when it exceeds the model context window.
user = None
class-attribute
instance-attribute
A unique identifier representing your end user.