Skip to content

Commit ab7edec

Browse files
committed
feat(wren-ai-service): add support for new GPT-5 models in configuration files
1 parent bf30acf commit ab7edec

7 files changed

Lines changed: 774 additions & 108 deletions

File tree

deployment/kustomizations/base/cm.yaml

Lines changed: 18 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -75,7 +75,24 @@ data:
7575
n: 1
7676
seed: 0
7777
temperature: 0
78-
78+
- model: gpt-5-nano-2025-08-07
79+
context_window_size: 1000000
80+
kwargs:
81+
max_tokens: 4096
82+
n: 1
83+
seed: 0
84+
- model: gpt-5-mini-2025-08-07
85+
context_window_size: 1000000
86+
kwargs:
87+
max_tokens: 4096
88+
n: 1
89+
seed: 0
90+
- model: gpt-5-2025-08-07
91+
context_window_size: 1000000
92+
kwargs:
93+
max_tokens: 4096
94+
n: 1
95+
seed: 0
7996
---
8097
type: embedder
8198
provider: litellm_embedder

docker/config.example.yaml

Lines changed: 19 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,25 @@ models:
2424
n: 1
2525
seed: 0
2626
temperature: 0
27-
27+
- model: gpt-5-nano-2025-08-07
28+
context_window_size: 1000000
29+
kwargs:
30+
max_tokens: 4096
31+
n: 1
32+
seed: 0
33+
- model: gpt-5-mini-2025-08-07
34+
context_window_size: 1000000
35+
kwargs:
36+
max_tokens: 4096
37+
n: 1
38+
seed: 0
39+
- model: gpt-5-2025-08-07
40+
context_window_size: 1000000
41+
kwargs:
42+
max_tokens: 4096
43+
n: 1
44+
seed: 0
45+
2846
---
2947
type: embedder
3048
provider: litellm_embedder

wren-ai-service/poetry.lock

Lines changed: 694 additions & 102 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

wren-ai-service/pyproject.toml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ package-mode = false
1010
[tool.poetry.dependencies]
1111
python = ">=3.12.*, <3.13"
1212
fastapi = "^0.115.2"
13-
uvicorn = {extras = ["standard"], version = "^0.30.1"}
13+
uvicorn = {extras = ["standard"], version = "^0.29.0"} # Litellm requires uvicorn < 0.30.0
1414
python-dotenv = "^1.0.1"
1515
haystack-ai = "==2.7.0"
1616
openai = "^1.40.0"
@@ -33,8 +33,8 @@ pydantic-settings = "^2.5.2"
3333
google-auth = "^2.35.0"
3434
tiktoken = "^0.8.0"
3535
jsonschema = "^4.23.0"
36-
litellm = "^1.70.0"
37-
boto3 = "^1.35.90"
36+
litellm = {extras = ["proxy"], version = "^1.75.2"}
37+
boto3 = "^1.34.34" # Litellm requires boto3 = 1.34.34
3838
qdrant-client = "==1.11.0"
3939

4040
[tool.poetry.group.dev.dependencies]

wren-ai-service/tools/config/config.example.yaml

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,24 @@ models:
2424
n: 1
2525
seed: 0
2626
temperature: 0
27+
- model: gpt-5-nano-2025-08-07
28+
context_window_size: 1000000
29+
kwargs:
30+
max_tokens: 4096
31+
n: 1
32+
seed: 0
33+
- model: gpt-5-mini-2025-08-07
34+
context_window_size: 1000000
35+
kwargs:
36+
max_tokens: 4096
37+
n: 1
38+
seed: 0
39+
- model: gpt-5-2025-08-07
40+
context_window_size: 1000000
41+
kwargs:
42+
max_tokens: 4096
43+
n: 1
44+
seed: 0
2745

2846
---
2947
type: embedder

wren-ai-service/tools/config/config.full.yaml

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,24 @@ models:
2424
n: 1
2525
seed: 0
2626
temperature: 0
27+
- model: gpt-5-nano-2025-08-07
28+
context_window_size: 1000000
29+
kwargs:
30+
max_tokens: 4096
31+
n: 1
32+
seed: 0
33+
- model: gpt-5-mini-2025-08-07
34+
context_window_size: 1000000
35+
kwargs:
36+
max_tokens: 4096
37+
n: 1
38+
seed: 0
39+
- model: gpt-5-2025-08-07
40+
context_window_size: 1000000
41+
kwargs:
42+
max_tokens: 4096
43+
n: 1
44+
seed: 0
2745

2846
---
2947
type: embedder

wren-launcher/commands/launch.go

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,7 @@ func askForGenerationModel() (string, error) {
109109

110110
prompt := promptui.Select{
111111
Label: "Select an OpenAI's generation model",
112-
Items: []string{"gpt-4.1", "gpt-4.1-mini", "gpt-4.1-nano"},
112+
Items: []string{"gpt-4.1", "gpt-4.1-mini", "gpt-4.1-nano", "gpt-5", "gpt-5-mini", "gpt-5-nano"},
113113
}
114114

115115
_, result, err := prompt.Run()
@@ -389,6 +389,9 @@ func getOpenaiGenerationModel() (string, bool) {
389389
"gpt-4.1": true,
390390
"gpt-4.1-mini": true,
391391
"gpt-4.1-nano": true,
392+
"gpt-5": true,
393+
"gpt-5-mini": true,
394+
"gpt-5-nano": true,
392395
}
393396
if !validModels[openaiGenerationModel] {
394397
pterm.Error.Println("Invalid generation model", openaiGenerationModel)

0 commit comments

Comments
 (0)