mirror of
https://github.com/bitnami/charts.git
synced 2026-04-02 15:27:08 +08:00
* [bitnami/kuberay] Release 1.4.14 updating components versions Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * Update CHANGELOG.md Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * Update CHANGELOG.md Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * [bitnami/kuberay] Release 1.4.14 updating components versions Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * Update CHANGELOG.md Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * [bitnami/kuberay] Release 1.4.14 updating components versions Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * Update CHANGELOG.md Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * Update CHANGELOG.md Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * Update CHANGELOG.md Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * [bitnami/kuberay] Increase resourcesPreset for cluster.head pod Signed-off-by: David Gomez <david.gomez@broadcom.com> --------- Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> Signed-off-by: David Gomez <david.gomez@broadcom.com> Co-authored-by: Fran Mulero <francisco-jose.mulero@broadcom.com> Co-authored-by: David Gomez <david.gomez@broadcom.com>
130 lines
3.7 KiB
YAML
130 lines
3.7 KiB
YAML
operator:
|
|
enabled: true
|
|
# Necessary to avoid conflicts in parallel deployments
|
|
watchAllNamespaces: false
|
|
containerPorts:
|
|
metrics: 8100
|
|
service:
|
|
ports:
|
|
metrics: 8104
|
|
serviceAccount:
|
|
create: true
|
|
automountServiceAccountToken: true
|
|
containerSecurityContext:
|
|
enabled: true
|
|
runAsUser: 1002
|
|
runAsGroup: 1002
|
|
runAsNonRoot: true
|
|
readOnlyRootFilesystem: true
|
|
allowPrivilegeEscalation: false
|
|
capabilities:
|
|
drop: ["ALL"]
|
|
podSecurityContext:
|
|
enabled: true
|
|
fsGroup: 1002
|
|
seccompProfile:
|
|
type: RuntimeDefault
|
|
metrics:
|
|
enabled: true
|
|
|
|
apiserver:
|
|
enabled: true
|
|
# Necessary to avoid conflicts in parallel deployments
|
|
watchAllNamespaces: false
|
|
service:
|
|
ports:
|
|
http: 8300
|
|
grpc: 8302
|
|
|
|
cluster:
|
|
enabled: true
|
|
service:
|
|
type: ClusterIP
|
|
worker:
|
|
groupSpecs:
|
|
- groupName: vib
|
|
head:
|
|
resourcesPreset: xlarge
|
|
|
|
extraDeploy:
|
|
# HACK: Kuberay does not allow setting a different service port from the container port, so we cannot set it at
|
|
# 80 because of the permissions. Instead, we create a mock service that points to the cluster head at port 80
|
|
- apiVersion: v1
|
|
kind: Service
|
|
metadata:
|
|
name: kuberay-cluster-head-vib-svc
|
|
labels:
|
|
app.kubernetes.io/component: cluster-head
|
|
spec:
|
|
type: LoadBalancer
|
|
selector:
|
|
app.kubernetes.io/component: cluster-head
|
|
ports:
|
|
- port: 80
|
|
name: http
|
|
targetPort: 8265
|
|
# HACK: The kuberay operator directly crates pod objects (not deployment or statefulsets) when creating RayCluster
|
|
# objects. Therefore we cannot run goss tests on the head node. As an alternative, we will create a job that submits
|
|
# a job to the RayCluster and we will check it in the Cypress tests
|
|
- apiVersion: batch/v1
|
|
kind: Job
|
|
metadata:
|
|
name: kuberay-vib-job
|
|
spec:
|
|
template:
|
|
spec:
|
|
restartPolicy: OnFailure
|
|
containers:
|
|
- name: job-runner
|
|
image: '{{ include "kuberay.ray.image" . }}'
|
|
securityContext:
|
|
runAsNonRoot: true
|
|
privileged: false
|
|
readOnlyRootFilesystem: true
|
|
allowPrivilegeEscalation: false
|
|
capabilities:
|
|
drop: ["ALL"]
|
|
seccompProfile:
|
|
type: "RuntimeDefault"
|
|
command:
|
|
- /bin/bash
|
|
args:
|
|
- -ec
|
|
- |
|
|
#!/bin/bash
|
|
|
|
set -o errexit
|
|
set -o nounset
|
|
set -o pipefail
|
|
|
|
. /opt/bitnami/scripts/libos.sh
|
|
|
|
# Set the endpoint URL
|
|
host="kuberay-cluster-head-svc"
|
|
port="8265"
|
|
|
|
kuberay_head_ready() {
|
|
# Test the TCP connection with a timeout
|
|
if timeout 5 bash -c "</dev/tcp/$host/$port"; then
|
|
return 0
|
|
else
|
|
return 1
|
|
fi
|
|
}
|
|
|
|
echo "0" > /tmp/ready
|
|
info "Waiting for the head instance"
|
|
if ! retry_while "kuberay_head_ready" 12 30; then
|
|
error "Could not connect to the head instance"
|
|
exit 1
|
|
else
|
|
info "Head ready! Running job"
|
|
ray job submit --address http://${host}:${port} -- python -c "import ray; ray.init(); print(ray.cluster_resources())"
|
|
fi
|
|
volumeMounts:
|
|
- name: tmp
|
|
mountPath: /tmp
|
|
volumes:
|
|
- name: tmp
|
|
emptyDir: {}
|