From 5e1cd0120f34670534ed4d596988a27cea1488f9 Mon Sep 17 00:00:00 2001 From: Pavlo Tkach <3469726+ptkach@users.noreply.github.com> Date: Thu, 28 Aug 2025 14:49:16 -0400 Subject: [PATCH] Adjust proxy resource allocation and update nomulus compute class (#2814) --- jetty/kubernetes/nomulus-backend.yaml | 7 ++++++- jetty/kubernetes/nomulus-frontend.yaml | 13 ++++++++++++- proxy/deploy-proxy-for-env.sh | 2 +- proxy/kubernetes/proxy-limit-range.yaml | 4 ++-- 4 files changed, 21 insertions(+), 5 deletions(-) diff --git a/jetty/kubernetes/nomulus-backend.yaml b/jetty/kubernetes/nomulus-backend.yaml index 303e8335a..68ee8c94f 100644 --- a/jetty/kubernetes/nomulus-backend.yaml +++ b/jetty/kubernetes/nomulus-backend.yaml @@ -15,7 +15,6 @@ spec: spec: serviceAccountName: nomulus nodeSelector: - cloud.google.com/compute-class: "Performance" cloud.google.com/machine-family: c4 containers: - name: backend @@ -25,9 +24,15 @@ spec: name: http resources: requests: + # explicit pod-slots 0 is required in order to downgrade node + # class from performance, which has implicit pod-slots 1 + cloud.google.com/pod-slots: 0 cpu: "500m" memory: "1Gi" limits: + # explicit pod-slots 0 is required in order to downgrade node + # class from performance, which has implicit pod-slots 1 + cloud.google.com/pod-slots: 0 cpu: "1000m" memory: "1.5Gi" args: [ENVIRONMENT] diff --git a/jetty/kubernetes/nomulus-frontend.yaml b/jetty/kubernetes/nomulus-frontend.yaml index 254be2459..648a90c42 100644 --- a/jetty/kubernetes/nomulus-frontend.yaml +++ b/jetty/kubernetes/nomulus-frontend.yaml @@ -15,7 +15,6 @@ spec: spec: serviceAccountName: nomulus nodeSelector: - cloud.google.com/compute-class: "Performance" cloud.google.com/machine-family: c4 containers: - name: frontend @@ -25,9 +24,15 @@ spec: name: http resources: requests: + # explicit pod-slots 0 is required in order to downgrade node + # class from performance, which has implicit pod-slots 1 + cloud.google.com/pod-slots: 0 cpu: "600m" memory: "1Gi" limits: + # explicit pod-slots 0 is required in order to downgrade node + # class from performance, which has implicit pod-slots 1 + cloud.google.com/pod-slots: 0 cpu: "1000m" memory: "2Gi" args: [ENVIRONMENT] @@ -53,9 +58,15 @@ spec: name: epp resources: requests: + # explicit pod-slots 0 is required in order to downgrade node + # class from performance, which has implicit pod-slots 1 + cloud.google.com/pod-slots: 0 cpu: "600m" memory: "512Mi" limits: + # explicit pod-slots 0 is required in order to downgrade node + # class from performance, which has implicit pod-slots 1 + cloud.google.com/pod-slots: 0 cpu: "1000m" memory: "1Gi" args: [--env, PROXY_ENV, --log, --local] diff --git a/proxy/deploy-proxy-for-env.sh b/proxy/deploy-proxy-for-env.sh index 3eeca05d0..80bb659c5 100755 --- a/proxy/deploy-proxy-for-env.sh +++ b/proxy/deploy-proxy-for-env.sh @@ -31,9 +31,9 @@ do echo "Updating cluster ${parts[0]} in zone ${parts[1]}..." gcloud container clusters get-credentials "${parts[0]}" \ --project "${project}" --zone "${parts[1]}" + kubectl apply -f "./kubernetes/proxy-limit-range.yaml" --force sed s/GCP_PROJECT/${project}/g "./kubernetes/proxy-deployment-${environment}.yaml" | \ kubectl apply -f - - kubectl apply -f "./kubernetes/proxy-limit-range.yaml" --force kubectl apply -f "./kubernetes/proxy-service.yaml" --force # Alpha does not have canary if [[ ${environment} != "alpha" ]]; then diff --git a/proxy/kubernetes/proxy-limit-range.yaml b/proxy/kubernetes/proxy-limit-range.yaml index 9c01bf1cd..fdd4b3efb 100644 --- a/proxy/kubernetes/proxy-limit-range.yaml +++ b/proxy/kubernetes/proxy-limit-range.yaml @@ -7,8 +7,8 @@ spec: limits: - type: Container default: - cpu: "300m" + cpu: "600m" memory: "512Mi" defaultRequest: - cpu: "100m" + cpu: "400m" memory: "350Mi"