Model Configuration
Overview
Section titled “Overview”Model configuration in any-agent is designed to be consistent across all supported frameworks. We use any-llm as the default model provider, which acts as a unified interface allowing you to use any language model from any provider with the same syntax.
Configuration Parameters
Section titled “Configuration Parameters”The model configuration is defined through several parameters in AgentConfig:
The model_id parameter selects which language model your agent will use. The format depends on the provider.
The model_args parameter allows you to pass additional arguments to the model, such as temperature, top_k, and other provider-specific parameters.
The api_base parameter allows you to specify a custom API endpoint. This is useful when:
- Using a local model server (e.g., Ollama, llama.cpp, llamafile)
- Routing through a proxy
- Using a self-hosted model endpoint
The api_key parameter allows you to explicitly specify an API key for authentication. By default, any-llm will automatically search for common environment variables (like OPENAI_API_KEY, ANTHROPIC_API_KEY, etc.).
See the AnyLLM Provider Documentation for the complete list of supported providers.