Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
19 changes: 18 additions & 1 deletion deployment/kustomizations/base/cm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,24 @@ data:
n: 1
seed: 0
temperature: 0

- model: gpt-5-nano-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0
- model: gpt-5-mini-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0
- model: gpt-5-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0
---
type: embedder
provider: litellm_embedder
Expand Down
18 changes: 18 additions & 0 deletions docker/config.example.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,24 @@ models:
n: 1
seed: 0
temperature: 0
- model: gpt-5-nano-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0
- model: gpt-5-mini-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0
- model: gpt-5-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0

---
type: embedder
Expand Down
796 changes: 694 additions & 102 deletions wren-ai-service/poetry.lock

Large diffs are not rendered by default.

6 changes: 3 additions & 3 deletions wren-ai-service/pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ package-mode = false
[tool.poetry.dependencies]
python = ">=3.12.*, <3.13"
fastapi = "^0.115.2"
uvicorn = {extras = ["standard"], version = "^0.30.1"}
uvicorn = {extras = ["standard"], version = "^0.29.0"} # Litellm requires uvicorn < 0.30.0
python-dotenv = "^1.0.1"
haystack-ai = "==2.7.0"
openai = "^1.40.0"
Expand All @@ -33,8 +33,8 @@ pydantic-settings = "^2.5.2"
google-auth = "^2.35.0"
tiktoken = "^0.8.0"
jsonschema = "^4.23.0"
litellm = "^1.70.0"
boto3 = "^1.35.90"
litellm = {extras = ["proxy"], version = "^1.75.2"}
boto3 = "^1.34.34" # Litellm requires boto3 = 1.34.34
qdrant-client = "==1.11.0"

[tool.poetry.group.dev.dependencies]
Expand Down
18 changes: 18 additions & 0 deletions wren-ai-service/tools/config/config.example.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,24 @@ models:
n: 1
seed: 0
temperature: 0
- model: gpt-5-nano-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0
- model: gpt-5-mini-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0
- model: gpt-5-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0

---
type: embedder
Expand Down
18 changes: 18 additions & 0 deletions wren-ai-service/tools/config/config.full.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,24 @@ models:
n: 1
seed: 0
temperature: 0
- model: gpt-5-nano-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0
- model: gpt-5-mini-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0
- model: gpt-5-2025-08-07
context_window_size: 380000
kwargs:
max_completion_tokens: 4096
n: 1
seed: 0

---
type: embedder
Expand Down
5 changes: 4 additions & 1 deletion wren-launcher/commands/launch.go
Original file line number Diff line number Diff line change
Expand Up @@ -109,7 +109,7 @@ func askForGenerationModel() (string, error) {

prompt := promptui.Select{
Label: "Select an OpenAI's generation model",
Items: []string{"gpt-4.1", "gpt-4.1-mini", "gpt-4.1-nano"},
Items: []string{"gpt-4.1", "gpt-4.1-mini", "gpt-4.1-nano", "gpt-5", "gpt-5-mini", "gpt-5-nano"},
}

_, result, err := prompt.Run()
Expand Down Expand Up @@ -389,6 +389,9 @@ func getOpenaiGenerationModel() (string, bool) {
"gpt-4.1": true,
"gpt-4.1-mini": true,
"gpt-4.1-nano": true,
"gpt-5": true,
"gpt-5-mini": true,
"gpt-5-nano": true,
}
if !validModels[openaiGenerationModel] {
pterm.Error.Println("Invalid generation model", openaiGenerationModel)
Expand Down
3 changes: 3 additions & 0 deletions wren-launcher/utils/docker.go
Original file line number Diff line number Diff line change
Expand Up @@ -34,6 +34,9 @@ var generationModelToModelName = map[string]string{
"gpt-4.1": "gpt-4.1-2025-04-14",
"gpt-4.1-mini": "gpt-4.1-mini-2025-04-14",
"gpt-4.1-nano": "gpt-4.1-nano-2025-04-14",
"gpt-5": "gpt-5-2025-08-07",
"gpt-5-mini": "gpt-5-mini-2025-08-07",
"gpt-5-nano": "gpt-5-nano-2025-08-07",
}

func replaceEnvFileContent(content string, projectDir string, openaiApiKey string, openAIGenerationModel string, hostPort int, aiPort int, userUUID string, telemetryEnabled bool, platform string, localStorage string) string {
Expand Down
Loading