Skip to content
Merged
16 changes: 8 additions & 8 deletions helm-chart/ray-cluster/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -88,9 +88,9 @@ helm uninstall raycluster
| head.containerEnv | list | `[]` | |
| head.envFrom | list | `[]` | envFrom to pass to head pod |
| head.resources.limits.cpu | string | `"1"` | |
| head.resources.limits.memory | string | `"2G"` | |
| head.resources.limits.memory | string | `"5Gi"` | |
| head.resources.requests.cpu | string | `"1"` | |
| head.resources.requests.memory | string | `"2G"` | |
| head.resources.requests.memory | string | `"5Gi"` | |
| head.resourceClaims | list | `[]` | ResourceClaims to allocate with the head pod |
| head.annotations | object | `{}` | Extra annotations for head pod |
| head.nodeSelector | object | `{}` | Node labels for head pod assignment |
Expand All @@ -111,7 +111,7 @@ helm uninstall raycluster
| worker.groupName | string | `"workergroup"` | The name of the workergroup |
| worker.replicas | int | `1` | The number of replicas for the worker pod |
| worker.minReplicas | int | `1` | The minimum number of replicas for the worker pod |
| worker.maxReplicas | int | `3` | The maximum number of replicas for the worker pod |
| worker.maxReplicas | int | `5` | The maximum number of replicas for the worker pod |
| worker.labels | object | `{}` | Labels for the worker pod |
| worker.serviceAccountName | string | `""` | |
| worker.restartPolicy | string | `""` | |
Expand All @@ -120,9 +120,9 @@ helm uninstall raycluster
| worker.containerEnv | list | `[]` | |
| worker.envFrom | list | `[]` | envFrom to pass to worker pod |
| worker.resources.limits.cpu | string | `"1"` | |
| worker.resources.limits.memory | string | `"1G"` | |
| worker.resources.limits.memory | string | `"1Gi"` | |
| worker.resources.requests.cpu | string | `"1"` | |
| worker.resources.requests.memory | string | `"1G"` | |
| worker.resources.requests.memory | string | `"1Gi"` | |
| worker.resourceClaims | list | `[]` | ResourceClaims to allocate with the worker pod |
| worker.annotations | object | `{}` | Extra annotations for worker pod |
| worker.nodeSelector | object | `{}` | Node labels for worker pod assignment |
Expand All @@ -142,17 +142,17 @@ helm uninstall raycluster
| additionalWorkerGroups.smallGroup.disabled | bool | `true` | |
| additionalWorkerGroups.smallGroup.replicas | int | `0` | The number of replicas for the additional worker pod |
| additionalWorkerGroups.smallGroup.minReplicas | int | `0` | The minimum number of replicas for the additional worker pod |
| additionalWorkerGroups.smallGroup.maxReplicas | int | `3` | The maximum number of replicas for the additional worker pod |
| additionalWorkerGroups.smallGroup.maxReplicas | int | `5` | The maximum number of replicas for the additional worker pod |
| additionalWorkerGroups.smallGroup.labels | object | `{}` | Labels for the additional worker pod |
| additionalWorkerGroups.smallGroup.serviceAccountName | string | `""` | |
| additionalWorkerGroups.smallGroup.restartPolicy | string | `""` | |
| additionalWorkerGroups.smallGroup.runtimeClassName | string | `""` | runtimeClassName for this additional worker group. Empty string means default runtime. |
| additionalWorkerGroups.smallGroup.containerEnv | list | `[]` | |
| additionalWorkerGroups.smallGroup.envFrom | list | `[]` | envFrom to pass to additional worker pod |
| additionalWorkerGroups.smallGroup.resources.limits.cpu | int | `1` | |
| additionalWorkerGroups.smallGroup.resources.limits.memory | string | `"1G"` | |
| additionalWorkerGroups.smallGroup.resources.limits.memory | string | `"1Gi"` | |
| additionalWorkerGroups.smallGroup.resources.requests.cpu | int | `1` | |
| additionalWorkerGroups.smallGroup.resources.requests.memory | string | `"1G"` | |
| additionalWorkerGroups.smallGroup.resources.requests.memory | string | `"1Gi"` | |
| additionalWorkerGroups.smallGroup.resourceClaims | list | `[]` | ResourceClaims to allocate with the additional worker pod |
| additionalWorkerGroups.smallGroup.annotations | object | `{}` | Extra annotations for additional worker pod |
| additionalWorkerGroups.smallGroup.nodeSelector | object | `{}` | Node labels for additional worker pod assignment |
Expand Down
16 changes: 8 additions & 8 deletions helm-chart/ray-cluster/values.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -128,10 +128,10 @@ head:
limits:
cpu: "1"
# To avoid out-of-memory issues, never allocate less than 2G memory for the Ray head.
memory: "2G"
memory: "5Gi"
requests:
cpu: "1"
memory: "2G"
memory: "5Gi"

# -- ResourceClaims to allocate with the head pod
resourceClaims: []
Expand Down Expand Up @@ -210,7 +210,7 @@ worker:
minReplicas: 1

# -- The maximum number of replicas for the worker pod
maxReplicas: 3
maxReplicas: 5

# -- Labels for the worker pod
labels: {}
Expand Down Expand Up @@ -249,10 +249,10 @@ worker:
resources:
limits:
cpu: "1"
memory: "1G"
memory: "1Gi"
requests:
cpu: "1"
memory: "1G"
memory: "1Gi"

# -- ResourceClaims to allocate with the worker pod
resourceClaims: []
Expand Down Expand Up @@ -324,7 +324,7 @@ additionalWorkerGroups:
minReplicas: 0

# -- The maximum number of replicas for the additional worker pod
maxReplicas: 3
maxReplicas: 5

# -- Labels for the additional worker pod
labels: {}
Expand Down Expand Up @@ -360,10 +360,10 @@ additionalWorkerGroups:
resources:
limits:
cpu: 1
memory: "1G"
memory: "1Gi"
requests:
cpu: 1
memory: "1G"
memory: "1Gi"

# -- ResourceClaims to allocate with the additional worker pod
resourceClaims: []
Expand Down
2 changes: 1 addition & 1 deletion kubectl-plugin/test/e2e/kubectl_ray_cluster_get_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,7 @@ var _ = Describe("Calling ray plugin `get` command", func() {
"2",
"0",
"0",
"3Gi",
"6Gi",
rayv1.RayClusterProvisioned,
rayv1.Ready,
},
Expand Down
2 changes: 1 addition & 1 deletion ray-operator/config/samples/ray-cluster.sample.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ spec:
memory: "5Gi"
requests:
cpu: "1"
memory: "2Gi"
memory: "5Gi"
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Btw, I think the e2e test output is failing since you increased the memory here:


[2026-03-16T15:43:22Z] Running Suite: Kubectl Ray e2e Test Suite - /workdir/kubectl-plugin/test/e2e
--
[2026-03-16T15:43:22Z] ============================================================================
[2026-03-16T15:43:22Z] Random Seed: 1773674078
[2026-03-16T15:43:22Z]
[2026-03-16T15:43:22Z] Will run 30 of 30 specs
[2026-03-16T15:43:22Z] ••••••••••••
[2026-03-16T15:43:22Z] ##############################
[2026-03-16T15:43:22Z] • [FAILED] [36.174 seconds]
[2026-03-16T15:43:22Z] Calling ray plugin `get` command [It] succeed in getting ray cluster information
[2026-03-16T15:43:22Z] /workdir/kubectl-plugin/test/e2e/kubectl_ray_cluster_get_test.go:28
[2026-03-16T15:43:22Z]
[2026-03-16T15:43:22Z]   [FAILED] Expected
[2026-03-16T15:43:22Z]       <string>: NAME                 NAMESPACE       DESIRED WORKERS   AVAILABLE WORKERS   CPUS   GPUS   TPUS   MEMORY   CONDITION               STATUS   AGE
[2026-03-16T15:43:22Z]       raycluster-kuberay   test-ns-usf5j   1                 1                   2      0      0      6Gi      RayClusterProvisioned   ready    25s
[2026-03-16T15:43:22Z]   to contain substring
[2026-03-16T15:43:22Z]       <string>: NAME                 NAMESPACE       DESIRED WORKERS   AVAILABLE WORKERS   CPUS   GPUS   TPUS   MEMORY   CONDITION               STATUS   AGE
[2026-03-16T15:43:22Z]       raycluster-kuberay   test-ns-usf5j   1                 1                   2      0      0      3Gi      RayClusterProvisioned   ready
[2026-03-16T15:43:22Z]   In [It] at: /workdir/kubectl-plugin/test/e2e/kubectl_ray_cluster_get_test.go:69 @ 03/16/26 15:28:44.638
[2026-03-16T15:43:22Z] ##############################
[2026-03-16T15:43:22Z] •••••••••••••••••
[2026-03-16T15:43:22Z]
[2026-03-16T15:43:22Z] Summarizing 1 Failure:
[2026-03-16T15:43:22Z]   [FAIL] Calling ray plugin `get` command [It] succeed in getting ray cluster information
[2026-03-16T15:43:22Z]   /workdir/kubectl-plugin/test/e2e/kubectl_ray_cluster_get_test.go:69
[2026-03-16T15:43:22Z]
[2026-03-16T15:43:22Z] Ran 30 of 30 Specs in 1724.141 seconds
[2026-03-16T15:43:22Z] FAIL! -- 29 Passed \| 1 Failed \| 0 Pending \| 0 Skipped
[2026-03-16T15:43:22Z] ### FAIL: TestKubectlRayCommand (1724.14s)
[2026-03-16T15:43:22Z] FAIL
[2026-03-16T15:43:22Z] FAIL	github.com/ray-project/kuberay/kubectl-plugin/test/e2e	1724.148s
[2026-03-16T15:43:22Z] FAIL
[2026-03-16T15:43:24Z] 🚨 Error: The command exited with status 1
[2026-03-16T15:43:24Z] user command error: The plugin docker command hook exited with status 1


Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

fixed, tks!

ports:
- containerPort: 6379
name: gcs-server
Expand Down
Loading