From 54f1ced244e91b356e8be610c86204ad61c9c37a Mon Sep 17 00:00:00 2001 From: Aaron Crickenberger Date: Fri, 7 Aug 2020 13:09:40 -0700 Subject: [PATCH 1/2] Migrate release-blocking ingress jobs to k8s-infra-prow-build Based on https://testgrid.k8s.io/sig-testing-canaries#gci-gce-ingress it appears these jobs do just fine with the default set of quotas available in the GCP projects in the default gce-project pool A followup change will remove the canary job once we have sufficient history/signal from the release-blocking jobs --- config/jobs/kubernetes/generated/generated.yaml | 8 ++++---- config/jobs/kubernetes/sig-network/sig-network-misc.yaml | 3 +-- releng/test_config.yaml | 2 +- 3 files changed, 6 insertions(+), 7 deletions(-) diff --git a/config/jobs/kubernetes/generated/generated.yaml b/config/jobs/kubernetes/generated/generated.yaml index 905474cdff6f..89d0d572e45e 100644 --- a/config/jobs/kubernetes/generated/generated.yaml +++ b/config/jobs/kubernetes/generated/generated.yaml @@ -1642,7 +1642,6 @@ periodics: - --gcp-zone=us-west1-b - --gcp-node-image=gci - --extract=ci/latest-1.19 - - --gcp-project-type=ingress-project - --timeout=150m - --test_args=--ginkgo.focus=\[Feature:Ingress\] --minStartupPods=8 env: @@ -1654,6 +1653,7 @@ periodics: limits: cpu: 1000m memory: 3Gi + cluster: k8s-infra-prow-build annotations: testgrid-tab-name: gce-cos-k8sbeta-ingress testgrid-dashboards: sig-release-1.19-blocking @@ -1866,7 +1866,6 @@ periodics: - --gcp-zone=us-west1-b - --gcp-node-image=gci - --extract=ci/latest-1.18 - - --gcp-project-type=ingress-project - --timeout=150m - --test_args=--ginkgo.focus=\[Feature:Ingress\] --minStartupPods=8 env: @@ -1878,6 +1877,7 @@ periodics: limits: cpu: 1000m memory: 3Gi + cluster: k8s-infra-prow-build annotations: testgrid-tab-name: gce-cos-k8sstable1-ingress testgrid-dashboards: sig-release-1.18-blocking @@ -2090,7 +2090,6 @@ periodics: - --gcp-zone=us-west1-b - --gcp-node-image=gci - --extract=ci/latest-1.17 - - --gcp-project-type=ingress-project - --timeout=150m - --test_args=--ginkgo.focus=\[Feature:Ingress\] --minStartupPods=8 env: @@ -2102,6 +2101,7 @@ periodics: limits: cpu: 1000m memory: 3Gi + cluster: k8s-infra-prow-build annotations: testgrid-tab-name: gce-cos-k8sstable2-ingress testgrid-dashboards: sig-release-1.17-blocking @@ -2278,7 +2278,6 @@ periodics: - --gcp-zone=us-west1-b - --gcp-node-image=gci - --extract=ci/latest-1.16 - - --gcp-project-type=ingress-project - --timeout=150m - --test_args=--ginkgo.focus=\[Feature:Ingress\] --minStartupPods=8 env: @@ -2290,6 +2289,7 @@ periodics: limits: cpu: 1000m memory: 3Gi + cluster: k8s-infra-prow-build annotations: testgrid-tab-name: gce-cos-k8sstable3-ingress testgrid-dashboards: sig-release-1.16-blocking diff --git a/config/jobs/kubernetes/sig-network/sig-network-misc.yaml b/config/jobs/kubernetes/sig-network/sig-network-misc.yaml index f8d4cd6f158d..166f99de59a6 100644 --- a/config/jobs/kubernetes/sig-network/sig-network-misc.yaml +++ b/config/jobs/kubernetes/sig-network/sig-network-misc.yaml @@ -248,6 +248,7 @@ periodics: - interval: 30m name: ci-kubernetes-e2e-gci-gce-ingress + cluster: k8s-infra-prow-build labels: preset-service-account: "true" preset-k8s-ssh: "true" @@ -263,7 +264,6 @@ periodics: - --env=KUBE_GCE_ENABLE_IP_ALIASES=true - --extract=ci/latest-fast - --gcp-node-image=gci - - --gcp-project-type=ingress-project - --gcp-zone=asia-southeast1-a - --provider=gce - --test_args=--ginkgo.focus=\[Feature:Ingress\]|\[Feature:NEG\] @@ -300,7 +300,6 @@ periodics: - --env=KUBE_GCE_ENABLE_IP_ALIASES=true - --extract=ci/latest-fast - --gcp-node-image=gci - - --gcp-project=k8s-infra-e2e-ingress-project - --gcp-zone=asia-southeast1-a - --provider=gce - --test_args=--ginkgo.focus=\[Feature:Ingress\]|\[Feature:NEG\] diff --git a/releng/test_config.yaml b/releng/test_config.yaml index cf8746282258..9a3344b3514c 100644 --- a/releng/test_config.yaml +++ b/releng/test_config.yaml @@ -566,9 +566,9 @@ testSuites: cluster: k8s-infra-prow-build ingress: args: - - --gcp-project-type=ingress-project - --timeout=150m - --test_args=--ginkgo.focus=\[Feature:Ingress\] --minStartupPods=8 + cluster: k8s-infra-prow-build reboot: args: - --timeout=180m From eaa75cfabd3a37f7910867af1e482b73b6682c04 Mon Sep 17 00:00:00 2001 From: Aaron Crickenberger Date: Fri, 7 Aug 2020 13:40:10 -0700 Subject: [PATCH 2/2] Raise static resource limits in jobs test I am halfway inclined to remove this as it's a poor approximation of resource usage, and I'm not sure what corrective action would be taken by someone other than just raising it (as I'm doing now). It's still the closest thing to a canary about resource usage not going way above what we would expect --- config/tests/jobs/jobs_test.go | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/config/tests/jobs/jobs_test.go b/config/tests/jobs/jobs_test.go index d954f8b432ab..41b01e9b66d1 100644 --- a/config/tests/jobs/jobs_test.go +++ b/config/tests/jobs/jobs_test.go @@ -1108,8 +1108,8 @@ func TestK8sInfraProwBuildJobsMustHavePodQOSGuaranteed(t *testing.T) { func TestK8sInfraProwBuildJobsMustNotExceedTotalCapacity(t *testing.T) { // k8s-infra-prow-build pool1 is 3-zonal 6-30 n1-highmem-8's maxLimit := coreapi.ResourceList{ - coreapi.ResourceCPU: resource.MustParse("288000m"), // 3 * 12 * 8 CPUs per n1-highmem-8 - coreapi.ResourceMemory: resource.MustParse("1872Gi"), // 3 * 12 * 52 Gi per n1-highmem-8 + coreapi.ResourceCPU: resource.MustParse("432"), // 3 * 18 * 8 CPUs per n1-highmem-8 + coreapi.ResourceMemory: resource.MustParse("2808Gi"), // 3 * 18 * 52 Gi per n1-highmem-8 } resourceNames := []coreapi.ResourceName{ coreapi.ResourceCPU,