Skip to content

Latest commit

 

History

History
56 lines (41 loc) · 2.28 KB

File metadata and controls

56 lines (41 loc) · 2.28 KB
layout default
title Chapter 3: LLM Provider Configuration
nav_order 3
parent Devika Tutorial

Chapter 3: LLM Provider Configuration

Welcome to Chapter 3: LLM Provider Configuration. In this part of Devika Tutorial: Open-Source Autonomous AI Software Engineer, you will build an intuitive mental model first, then move into concrete implementation details and practical production tradeoffs.

This chapter covers how to configure Claude 3, GPT-4, Gemini, Mistral, Groq, and local Ollama models in Devika's config.toml and how to select the right provider for each agent role.

Learning Goals

  • configure API keys and model identifiers for every supported LLM provider
  • understand Devika's model selection mechanism and how to switch providers per project
  • evaluate the cost, latency, and quality tradeoffs across providers for autonomous coding tasks
  • configure Ollama for fully offline, local LLM operation without external API keys

Fast Start Checklist

  1. open config.toml and locate the [API_KEYS] and [API_MODELS] sections
  2. add your API key for at least one cloud provider (Claude, OpenAI, Google, Mistral, or Groq)
  3. set the model name for each provider section to a currently available model identifier
  4. optionally install and start Ollama with a code-capable model for local operation

Source References

Summary

You now know how to configure any of Devika's supported LLM providers, select the right model for each use case, and operate Devika in fully local mode using Ollama.

Next: Chapter 4: Task Planning and Code Generation

How These Components Connect

flowchart TD
    A[config.yaml] --> B{Provider selection}
    B -->|Claude| C[ANTHROPIC_API_KEY]
    B -->|GPT-4| D[OPENAI_API_KEY]
    B -->|Gemini| E[GEMINI_API_KEY]
    B -->|Ollama| F[Local endpoint]
    C --> G[LLM abstraction layer]
    D --> G
    E --> G
    F --> G
    G --> H[Agent pipeline]
Loading