Skip to content

Commit e4e4c9f

Browse files
yichieh-lucyyeh
andauthored
feat(wren-ai-service): add support for new GPT-5 models in configuration files (ai-env-changed) (#1875)
Co-authored-by: ChihYu Yeh <[email protected]>
1 parent 15a5ddf commit e4e4c9f

8 files changed

Lines changed: 776 additions & 107 deletions

File tree

deployment/kustomizations/base/cm.yaml

Lines changed: 18 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -75,7 +75,24 @@ data:
7575
n: 1
7676
seed: 0
7777
temperature: 0
78-
78+
- model: gpt-5-nano-2025-08-07
79+
context_window_size: 380000
80+
kwargs:
81+
max_completion_tokens: 4096
82+
n: 1
83+
seed: 0
84+
- model: gpt-5-mini-2025-08-07
85+
context_window_size: 380000
86+
kwargs:
87+
max_completion_tokens: 4096
88+
n: 1
89+
seed: 0
90+
- model: gpt-5-2025-08-07
91+
context_window_size: 380000
92+
kwargs:
93+
max_completion_tokens: 4096
94+
n: 1
95+
seed: 0
7996
---
8097
type: embedder
8198
provider: litellm_embedder

docker/config.example.yaml

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,24 @@ models:
2424
n: 1
2525
seed: 0
2626
temperature: 0
27+
- model: gpt-5-nano-2025-08-07
28+
context_window_size: 380000
29+
kwargs:
30+
max_completion_tokens: 4096
31+
n: 1
32+
seed: 0
33+
- model: gpt-5-mini-2025-08-07
34+
context_window_size: 380000
35+
kwargs:
36+
max_completion_tokens: 4096
37+
n: 1
38+
seed: 0
39+
- model: gpt-5-2025-08-07
40+
context_window_size: 380000
41+
kwargs:
42+
max_completion_tokens: 4096
43+
n: 1
44+
seed: 0
2745

2846
---
2947
type: embedder

wren-ai-service/poetry.lock

Lines changed: 694 additions & 102 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

wren-ai-service/pyproject.toml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ package-mode = false
1010
[tool.poetry.dependencies]
1111
python = ">=3.12.*, <3.13"
1212
fastapi = "^0.115.2"
13-
uvicorn = {extras = ["standard"], version = "^0.30.1"}
13+
uvicorn = {extras = ["standard"], version = "^0.29.0"} # Litellm requires uvicorn < 0.30.0
1414
python-dotenv = "^1.0.1"
1515
haystack-ai = "==2.7.0"
1616
openai = "^1.40.0"
@@ -33,8 +33,8 @@ pydantic-settings = "^2.5.2"
3333
google-auth = "^2.35.0"
3434
tiktoken = "^0.8.0"
3535
jsonschema = "^4.23.0"
36-
litellm = "^1.70.0"
37-
boto3 = "^1.35.90"
36+
litellm = {extras = ["proxy"], version = "^1.75.2"}
37+
boto3 = "^1.34.34" # Litellm requires boto3 = 1.34.34
3838
qdrant-client = "==1.11.0"
3939

4040
[tool.poetry.group.dev.dependencies]

wren-ai-service/tools/config/config.example.yaml

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,24 @@ models:
2424
n: 1
2525
seed: 0
2626
temperature: 0
27+
- model: gpt-5-nano-2025-08-07
28+
context_window_size: 380000
29+
kwargs:
30+
max_completion_tokens: 4096
31+
n: 1
32+
seed: 0
33+
- model: gpt-5-mini-2025-08-07
34+
context_window_size: 380000
35+
kwargs:
36+
max_completion_tokens: 4096
37+
n: 1
38+
seed: 0
39+
- model: gpt-5-2025-08-07
40+
context_window_size: 380000
41+
kwargs:
42+
max_completion_tokens: 4096
43+
n: 1
44+
seed: 0
2745

2846
---
2947
type: embedder

wren-ai-service/tools/config/config.full.yaml

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,6 +24,24 @@ models:
2424
n: 1
2525
seed: 0
2626
temperature: 0
27+
- model: gpt-5-nano-2025-08-07
28+
context_window_size: 380000
29+
kwargs:
30+
max_completion_tokens: 4096
31+
n: 1
32+
seed: 0
33+
- model: gpt-5-mini-2025-08-07
34+
context_window_size: 380000
35+
kwargs:
36+
max_completion_tokens: 4096
37+
n: 1
38+
seed: 0
39+
- model: gpt-5-2025-08-07
40+
context_window_size: 380000
41+
kwargs:
42+
max_completion_tokens: 4096
43+
n: 1
44+
seed: 0
2745

2846
---
2947
type: embedder

wren-launcher/commands/launch.go

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,7 @@ func askForGenerationModel() (string, error) {
109109

110110
prompt := promptui.Select{
111111
Label: "Select an OpenAI's generation model",
112-
Items: []string{"gpt-4.1", "gpt-4.1-mini", "gpt-4.1-nano"},
112+
Items: []string{"gpt-4.1", "gpt-4.1-mini", "gpt-4.1-nano", "gpt-5", "gpt-5-mini", "gpt-5-nano"},
113113
}
114114

115115
_, result, err := prompt.Run()
@@ -389,6 +389,9 @@ func getOpenaiGenerationModel() (string, bool) {
389389
"gpt-4.1": true,
390390
"gpt-4.1-mini": true,
391391
"gpt-4.1-nano": true,
392+
"gpt-5": true,
393+
"gpt-5-mini": true,
394+
"gpt-5-nano": true,
392395
}
393396
if !validModels[openaiGenerationModel] {
394397
pterm.Error.Println("Invalid generation model", openaiGenerationModel)

wren-launcher/utils/docker.go

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -34,6 +34,9 @@ var generationModelToModelName = map[string]string{
3434
"gpt-4.1": "gpt-4.1-2025-04-14",
3535
"gpt-4.1-mini": "gpt-4.1-mini-2025-04-14",
3636
"gpt-4.1-nano": "gpt-4.1-nano-2025-04-14",
37+
"gpt-5": "gpt-5-2025-08-07",
38+
"gpt-5-mini": "gpt-5-mini-2025-08-07",
39+
"gpt-5-nano": "gpt-5-nano-2025-08-07",
3740
}
3841

3942
func replaceEnvFileContent(content string, projectDir string, openaiApiKey string, openAIGenerationModel string, hostPort int, aiPort int, userUUID string, telemetryEnabled bool, platform string, localStorage string) string {

0 commit comments

Comments
 (0)