| layout | default |
|---|---|
| title | Chapter 3: LLM Provider Configuration |
| nav_order | 3 |
| parent | Devika Tutorial |
Welcome to Chapter 3: LLM Provider Configuration. In this part of Devika Tutorial: Open-Source Autonomous AI Software Engineer, you will build an intuitive mental model first, then move into concrete implementation details and practical production tradeoffs.
This chapter covers how to configure Claude 3, GPT-4, Gemini, Mistral, Groq, and local Ollama models in Devika's config.toml and how to select the right provider for each agent role.
- configure API keys and model identifiers for every supported LLM provider
- understand Devika's model selection mechanism and how to switch providers per project
- evaluate the cost, latency, and quality tradeoffs across providers for autonomous coding tasks
- configure Ollama for fully offline, local LLM operation without external API keys
- open
config.tomland locate the[API_KEYS]and[API_MODELS]sections - add your API key for at least one cloud provider (Claude, OpenAI, Google, Mistral, or Groq)
- set the model name for each provider section to a currently available model identifier
- optionally install and start Ollama with a code-capable model for local operation
You now know how to configure any of Devika's supported LLM providers, select the right model for each use case, and operate Devika in fully local mode using Ollama.
Next: Chapter 4: Task Planning and Code Generation
flowchart TD
A[config.yaml] --> B{Provider selection}
B -->|Claude| C[ANTHROPIC_API_KEY]
B -->|GPT-4| D[OPENAI_API_KEY]
B -->|Gemini| E[GEMINI_API_KEY]
B -->|Ollama| F[Local endpoint]
C --> G[LLM abstraction layer]
D --> G
E --> G
F --> G
G --> H[Agent pipeline]