Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

K8s Prow is limited to 1k concurrent pods, or podgc will fight with sinker #11594

Closed
mm4tt opened this issue Mar 4, 2019 · 74 comments
Closed
Assignees
Labels
kind/bug Categorizes issue or PR as related to a bug. priority/important-soon Must be staffed and worked on either currently, or very soon, ideally in time for the next release. sig/scalability Categorizes an issue or PR as relevant to SIG Scalability. sig/testing Categorizes an issue or PR as relevant to SIG Testing.

Comments

@mm4tt
Copy link
Contributor

mm4tt commented Mar 4, 2019

Our scalability tests started behaving strangely over the weekend. The prow jobs running tests are scheduled when they shouldn't.

Example:
name: ci-kubernetes-e2e-gke-large-performance-regional
Config:
oqm1ekq6cvr 1

Test should be run once a week, but it has been scheduled 4 times over last weekend.
qtykinscgey

There are other jobs behaving similarly to this one, i.e. they are scheduled and run when they shouldn't be run.

This is wreaking havoc in our scalability tests.
Due to quota issues, the tests share the same gcp projects. Now, because they're run when shouldn't, they started interfering with each other causing multiple tests to fail.

@mm4tt mm4tt added the kind/bug Categorizes issue or PR as related to a bug. label Mar 4, 2019
@mm4tt
Copy link
Contributor Author

mm4tt commented Mar 4, 2019

/priority critical-urgent

@k8s-ci-robot k8s-ci-robot added the priority/critical-urgent Highest priority. Must be actively worked on as someone's top priority right now. label Mar 4, 2019
@mm4tt
Copy link
Contributor Author

mm4tt commented Mar 4, 2019

@fejta, could you take a look or reassign?
/assign @fejta

@mm4tt
Copy link
Contributor Author

mm4tt commented Mar 4, 2019

/sig testing

@k8s-ci-robot k8s-ci-robot added the sig/testing Categorizes an issue or PR as relevant to SIG Testing. label Mar 4, 2019
@stevekuznetsov
Copy link
Contributor

@mm4tt FYI the best place to escalate something like this is in #testing-ops on Slack, pinging @test-infra-oncall. I've shot a message over here

@spiffxp
Copy link
Member

spiffxp commented Mar 4, 2019

/milestone v1.14
/unassign @fejta
not available at the moment
/assign @amwat
as 1.14 test-infra lead, and currently on-call per: go.k8s.io/oncall

@k8s-ci-robot k8s-ci-robot added this to the v1.14 milestone Mar 4, 2019
@BenTheElder
Copy link
Member

There are other jobs behaving similarly to this one, i.e. they are scheduled and run when they shouldn't be run.

could you list them?

Strangely https://prow.k8s.io/?job=ci-kubernetes-e2e-gke-large-performance-regional has one entry at Mar 03 00:01:39

https://prow.k8s.io/rerun?prowjob=931c4290-3d8a-11e9-9c9a-0a580a6c0e78

apiVersion: prow.k8s.io/v1
kind: ProwJob
metadata:
  annotations:
    prow.k8s.io/job: ci-kubernetes-e2e-gke-large-performance-regional
  creationTimestamp: null
  labels:
    created-by-prow: "true"
    preset-k8s-ssh: "true"
    preset-service-account: "true"
    prow.k8s.io/id: 931c4290-3d8a-11e9-9c9a-0a580a6c0e78
    prow.k8s.io/job: ci-kubernetes-e2e-gke-large-performance-regional
    prow.k8s.io/type: periodic
  name: e89951f0-3e98-11e9-b844-0a580a6c0923
spec:
  agent: kubernetes
  cluster: default
  job: ci-kubernetes-e2e-gke-large-performance-regional
  namespace: test-pods
  pod_spec:
    containers:
    - args:
      - --timeout=600
      - --repo=k8s.io/kubernetes=master
      - --repo=k8s.io/perf-tests=master
      - --root=/go/src
      - --scenario=kubernetes_e2e
      - --
      - --cluster=gke-regional-cluster
      - --deployment=gke
      - --extract=ci/latest-1.13
      - --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging
      - --gcp-node-image=gci
      - --gcp-project=kubernetes-scale
      - --gcp-region=us-east1
      - --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork
        name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19
      - --gke-environment=staging
      - --gke-node-locations=us-east1-b
      - --gke-shape={"default":{"Nodes":1999,"MachineType":"n1-standard-1"},"heapster-pool":{"Nodes":1,"MachineType":"n1-standard-8"}}
      - --provider=gke
      - --test=false
      - --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh
      - --test-cmd-args=cluster-loader2
      - --test-cmd-args=--nodes=2000
      - --test-cmd-args=--provider=gke
      - --test-cmd-args=--report-dir=/workspace/_artifacts
      - --test-cmd-args=--testconfig=testing/density/config.yaml
      - --test-cmd-args=--testconfig=testing/load/config.yaml
      - --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml
      - --test-cmd-name=ClusterLoaderV2
      - --timeout=570m
      - --use-logexporter
      env:
      - name: GOOGLE_APPLICATION_CREDENTIALS
        value: /etc/service-account/service-account.json
      - name: E2E_GOOGLE_APPLICATION_CREDENTIALS
        value: /etc/service-account/service-account.json
      - name: USER
        value: prow
      - name: JENKINS_GCE_SSH_PRIVATE_KEY_FILE
        value: /etc/ssh-key-secret/ssh-private
      - name: JENKINS_GCE_SSH_PUBLIC_KEY_FILE
        value: /etc/ssh-key-secret/ssh-public
      image: gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master
      name: ""
      resources:
        requests:
          cpu: "6"
          memory: 16Gi
      volumeMounts:
      - mountPath: /etc/service-account
        name: service
        readOnly: true
      - mountPath: /etc/ssh-key-secret
        name: ssh
        readOnly: true
    volumes:
    - name: service
      secret:
        secretName: service-account
    - name: ssh
      secret:
        defaultMode: 256
        secretName: ssh-key-secret
  type: periodic
status:
  startTime: "2019-03-04T16:16:46Z"
  state: triggered

https://testgrid.k8s.io/sig-scalability-gke#gke-large-performance-regional

@stevekuznetsov
Copy link
Contributor

@BenTheElder I recently added more logging to horologium to discern why a job was triggered -- what are the logs saying there?

@BenTheElder
Copy link
Member

the logs for horologium don't seem to show much so far (stackdriver export for horologium pod, text:ci-kubernetes-e2e-gke-large-performance-regional, back to 3/2/19 2:00:00 AM)

[
 {
   "insertId": "1ulicp5fepvsxu",
   "jsonPayload": {
     "level": "info",
     "should-trigger": true,
     "name": "931c4290-3d8a-11e9-9c9a-0a580a6c0e78",
     "component": "horologium",
     "type": "periodic",
     "job": "ci-kubernetes-e2e-gke-large-performance-regional",
     "msg": "Triggering new run of cron periodic.",
     "previous-found": true
   },
   "resource": {
     "type": "container",
     "labels": {
       "zone": "us-central1-f",
       "pod_id": "horologium-78fb7b98f8-tn8sc",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "horologium",
       "namespace_id": "default",
       "instance_id": "5347705516640603225"
     }
   },
   "timestamp": "2019-03-03T08:01:39Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "horologium-78fb7b98f8-tn8sc",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-blkqd"
   },
   "logName": "projects/k8s-prow/logs/horologium",
   "receiveTimestamp": "2019-03-03T08:01:45.470497268Z"
 },
 {
   "insertId": "15wtu9nfepwm0g",
   "jsonPayload": {
     "component": "horologium",
     "msg": "Triggering cron job ci-kubernetes-e2e-gke-large-performance-regional.",
     "client": "cron",
     "level": "info"
   },
   "resource": {
     "type": "container",
     "labels": {
       "zone": "us-central1-f",
       "pod_id": "horologium-78fb7b98f8-tn8sc",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "horologium",
       "namespace_id": "default",
       "instance_id": "5347705516640603225"
     }
   },
   "timestamp": "2019-03-03T08:01:00Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "horologium-78fb7b98f8-tn8sc",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-blkqd"
   },
   "logName": "projects/k8s-prow/logs/horologium",
   "receiveTimestamp": "2019-03-03T08:01:06.402048467Z"
 }
]

@BenTheElder
Copy link
Member

plank for text:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 🤔

[
 {
   "insertId": "a85i6flj274m",
   "jsonPayload": {
     "component": "plank",
     "msg": "ReplaceProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {931c4290-3d8a-11e9-9c9a-0a580a6c0e78  default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/931c4290-3d8a-11e9-9c9a-0a580a6c0e78 933a4706-3d8a-11e9-898b-42010a80003a 189614343 1 2019-03-03 08:01:39 +0000 UTC <nil> <nil> map[created-by-prow:true preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional prow.k8s.io/type:periodic] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {periodic kubernetes default test-pods ci-kubernetes-e2e-gke-large-performance-regional <nil> [] false   0 false &PodSpec{Volumes:[{service {nil nil nil nil nil SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{ gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}],RestartPolicy:,TerminationGracePeriodSeconds:nil,ActiveDeadlineSeconds:nil,DNSPolicy:,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:nil,ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,} <nil> <nil>} {2019-03-03 08:01:39 +0000 UTC <nil> pending Job triggered. https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-large-performance-regional/165/ 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 165  map[github-reporter:pending]}})",
     "client": "kube",
     "level": "debug"
   },
   "resource": {
     "type": "container",
     "labels": {
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-04T13:08:23Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-04T13:08:28.783717431Z"
 },
 {
   "insertId": "a85i6flj274l",
   "jsonPayload": {
     "client": "kube",
     "level": "debug",
     "component": "plank",
     "msg": "GetProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78)"
   },
   "resource": {
     "type": "container",
     "labels": {
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-04T13:08:23Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-04T13:08:28.783717431Z"
 },
 {
   "insertId": "a85i6flj2748",
   "jsonPayload": {
     "msg": "ReplaceProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {931c4290-3d8a-11e9-9c9a-0a580a6c0e78  default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/931c4290-3d8a-11e9-9c9a-0a580a6c0e78 933a4706-3d8a-11e9-898b-42010a80003a 189487735 1 2019-03-03 08:01:39 +0000 UTC <nil> <nil> map[created-by-prow:true preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional prow.k8s.io/type:periodic] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {periodic kubernetes default test-pods ci-kubernetes-e2e-gke-large-performance-regional <nil> [] false   0 false &PodSpec{Volumes:[{service {nil nil nil nil nil SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{ gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}],RestartPolicy:,TerminationGracePeriodSeconds:nil,ActiveDeadlineSeconds:nil,DNSPolicy:,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:nil,ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,} <nil> <nil>} {2019-03-03 08:01:39 +0000 UTC <nil> pending Job triggered. https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-large-performance-regional/165/ 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 165  map[github-reporter:pending]}})",
     "client": "kube",
     "level": "debug",
     "component": "plank"
   },
   "resource": {
     "type": "container",
     "labels": {
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921",
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow"
     }
   },
   "timestamp": "2019-03-04T13:08:23Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-04T13:08:28.783717431Z"
 },
 {
   "insertId": "a85i6flj2747",
   "jsonPayload": {
     "job": "ci-kubernetes-e2e-gke-large-performance-regional",
     "msg": "Pod is missing, starting a new pod",
     "level": "info",
     "name": "931c4290-3d8a-11e9-9c9a-0a580a6c0e78",
     "type": "periodic",
     "component": "plank"
   },
   "resource": {
     "type": "container",
     "labels": {
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-04T13:08:23Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-04T13:08:28.783717431Z"
 },
 {
   "insertId": "a85i6flj2746",
   "jsonPayload": {
     "component": "plank",
     "msg": "CreatePod({{ } {931c4290-3d8a-11e9-9c9a-0a580a6c0e78      0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 created-by-prow:true prow.k8s.io/type:periodic prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional preset-k8s-ssh:true preset-service-account:true] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {[{service {nil nil nil nil nil &SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}] [] [{test gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil} {BUILD_ID 165 nil} {BUILD_NUMBER 165 nil} {JOB_NAME ci-kubernetes-e2e-gke-large-performance-regional nil} {JOB_SPEC {\"type\":\"periodic\",\"job\":\"ci-kubernetes-e2e-gke-large-performance-regional\",\"buildid\":\"165\",\"prowjobid\":\"931c4290-3d8a-11e9-9c9a-0a580a6c0e78\"} nil} {JOB_TYPE periodic nil} {PROW_JOB_ID 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}] Never <nil> <nil>  map[]   0xc02e0680b8  false false false <nil> nil []   nil  [] []  <nil> nil [] <nil>} { []      <nil> [] [] }})",
     "client": "kube",
     "level": "debug"
   },
   "resource": {
     "type": "container",
     "labels": {
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-04T13:08:23Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-04T13:08:28.783717431Z"
 },
 {
   "insertId": "1vzsmovfk87wnw",
   "jsonPayload": {
     "component": "plank",
     "msg": "ReplaceProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {931c4290-3d8a-11e9-9c9a-0a580a6c0e78  default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/931c4290-3d8a-11e9-9c9a-0a580a6c0e78 933a4706-3d8a-11e9-898b-42010a80003a 189487735 1 2019-03-03 08:01:39 +0000 UTC <nil> <nil> map[created-by-prow:true preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional prow.k8s.io/type:periodic] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {periodic kubernetes default test-pods ci-kubernetes-e2e-gke-large-performance-regional <nil> [] false   0 false &PodSpec{Volumes:[{service {nil nil nil nil nil SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{ gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}],RestartPolicy:,TerminationGracePeriodSeconds:nil,ActiveDeadlineSeconds:nil,DNSPolicy:,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:nil,ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,} <nil> <nil>} {2019-03-03 08:01:39 +0000 UTC <nil> pending Job triggered. https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-large-performance-regional/164/ 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 164  map[github-reporter:pending]}})",
     "client": "kube",
     "level": "debug"
   },
   "resource": {
     "type": "container",
     "labels": {
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow"
     }
   },
   "timestamp": "2019-03-04T05:43:23Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-04T05:43:28.556197331Z"
 },
 {
   "insertId": "1vzsmovfk87wnv",
   "jsonPayload": {
     "component": "plank",
     "msg": "GetProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78)",
     "client": "kube",
     "level": "debug"
   },
   "resource": {
     "type": "container",
     "labels": {
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow"
     }
   },
   "timestamp": "2019-03-04T05:43:23Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-04T05:43:28.556197331Z"
 },
 {
   "insertId": "1vzsmovfk87wnl",
   "jsonPayload": {
     "component": "plank",
     "msg": "ReplaceProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {931c4290-3d8a-11e9-9c9a-0a580a6c0e78  default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/931c4290-3d8a-11e9-9c9a-0a580a6c0e78 933a4706-3d8a-11e9-898b-42010a80003a 189363471 1 2019-03-03 08:01:39 +0000 UTC <nil> <nil> map[created-by-prow:true preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional prow.k8s.io/type:periodic] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {periodic kubernetes default test-pods ci-kubernetes-e2e-gke-large-performance-regional <nil> [] false   0 false &PodSpec{Volumes:[{service {nil nil nil nil nil SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{ gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}],RestartPolicy:,TerminationGracePeriodSeconds:nil,ActiveDeadlineSeconds:nil,DNSPolicy:,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:nil,ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,} <nil> <nil>} {2019-03-03 08:01:39 +0000 UTC <nil> pending Job triggered. https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-large-performance-regional/164/ 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 164  map[github-reporter:pending]}})",
     "client": "kube",
     "level": "debug"
   },
   "resource": {
     "type": "container",
     "labels": {
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921",
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow"
     }
   },
   "timestamp": "2019-03-04T05:43:23Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-04T05:43:28.556197331Z"
 },
 {
   "insertId": "1vzsmovfk87wnk",
   "jsonPayload": {
     "msg": "Pod is missing, starting a new pod",
     "job": "ci-kubernetes-e2e-gke-large-performance-regional",
     "level": "info",
     "name": "931c4290-3d8a-11e9-9c9a-0a580a6c0e78",
     "type": "periodic",
     "component": "plank"
   },
   "resource": {
     "type": "container",
     "labels": {
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-04T05:43:23Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-04T05:43:28.556197331Z"
 },
 {
   "insertId": "1vzsmovfk87wnj",
   "jsonPayload": {
     "msg": "CreatePod({{ } {931c4290-3d8a-11e9-9c9a-0a580a6c0e78      0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 created-by-prow:true prow.k8s.io/type:periodic prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {[{service {nil nil nil nil nil &SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}] [] [{test gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil} {BUILD_ID 164 nil} {BUILD_NUMBER 164 nil} {JOB_NAME ci-kubernetes-e2e-gke-large-performance-regional nil} {JOB_SPEC {\"type\":\"periodic\",\"job\":\"ci-kubernetes-e2e-gke-large-performance-regional\",\"buildid\":\"164\",\"prowjobid\":\"931c4290-3d8a-11e9-9c9a-0a580a6c0e78\"} nil} {JOB_TYPE periodic nil} {PROW_JOB_ID 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}] Never <nil> <nil>  map[]   0xc010a48048  false false false <nil> nil []   nil  [] []  <nil> nil [] <nil>} { []      <nil> [] [] }})",
     "client": "kube",
     "level": "debug",
     "component": "plank"
   },
   "resource": {
     "type": "container",
     "labels": {
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-04T05:43:23Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-04T05:43:28.556197331Z"
 },
 {
   "insertId": "10cniqefl3qiev",
   "jsonPayload": {
     "client": "kube",
     "level": "debug",
     "component": "plank",
     "msg": "ReplaceProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {931c4290-3d8a-11e9-9c9a-0a580a6c0e78  default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/931c4290-3d8a-11e9-9c9a-0a580a6c0e78 933a4706-3d8a-11e9-898b-42010a80003a 189363471 1 2019-03-03 08:01:39 +0000 UTC <nil> <nil> map[prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional prow.k8s.io/type:periodic created-by-prow:true preset-k8s-ssh:true preset-service-account:true] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {periodic kubernetes default test-pods ci-kubernetes-e2e-gke-large-performance-regional <nil> [] false   0 false &PodSpec{Volumes:[{service {nil nil nil nil nil SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{ gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}],RestartPolicy:,TerminationGracePeriodSeconds:nil,ActiveDeadlineSeconds:nil,DNSPolicy:,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:nil,ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,} <nil> <nil>} {2019-03-03 08:01:39 +0000 UTC <nil> pending Job triggered. https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-large-performance-regional/163/ 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 163  map[github-reporter:pending]}})"
   },
   "resource": {
     "type": "container",
     "labels": {
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T22:26:54Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T22:27:01.550080143Z"
 },
 {
   "insertId": "10cniqefl3qieu",
   "jsonPayload": {
     "client": "kube",
     "level": "debug",
     "component": "plank",
     "msg": "GetProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78)"
   },
   "resource": {
     "type": "container",
     "labels": {
       "namespace_id": "default",
       "instance_id": "7000980459144515921",
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank"
     }
   },
   "timestamp": "2019-03-03T22:26:54Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T22:27:01.550080143Z"
 },
 {
   "insertId": "10cniqefl3qiee",
   "jsonPayload": {
     "client": "kube",
     "level": "debug",
     "component": "plank",
     "msg": "ReplaceProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {931c4290-3d8a-11e9-9c9a-0a580a6c0e78  default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/931c4290-3d8a-11e9-9c9a-0a580a6c0e78 933a4706-3d8a-11e9-898b-42010a80003a 189240028 1 2019-03-03 08:01:39 +0000 UTC <nil> <nil> map[prow.k8s.io/type:periodic created-by-prow:true preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {periodic kubernetes default test-pods ci-kubernetes-e2e-gke-large-performance-regional <nil> [] false   0 false &PodSpec{Volumes:[{service {nil nil nil nil nil SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{ gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil}] {map[] map[memory:{{17179869184 0} {<nil>}  BinarySI} cpu:{{6 0} {<nil>} 6 DecimalSI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}],RestartPolicy:,TerminationGracePeriodSeconds:nil,ActiveDeadlineSeconds:nil,DNSPolicy:,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:nil,ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,} <nil> <nil>} {2019-03-03 08:01:39 +0000 UTC <nil> pending Job triggered. https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-large-performance-regional/163/ 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 163  map[github-reporter:pending]}})"
   },
   "resource": {
     "type": "container",
     "labels": {
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T22:26:54Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T22:27:01.550080143Z"
 },
 {
   "insertId": "10cniqefl3qied",
   "jsonPayload": {
     "msg": "Pod is missing, starting a new pod",
     "job": "ci-kubernetes-e2e-gke-large-performance-regional",
     "level": "info",
     "name": "931c4290-3d8a-11e9-9c9a-0a580a6c0e78",
     "type": "periodic",
     "component": "plank"
   },
   "resource": {
     "type": "container",
     "labels": {
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T22:26:54Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T22:27:01.550080143Z"
 },
 {
   "insertId": "10cniqefl3qiec",
   "jsonPayload": {
     "msg": "CreatePod({{ } {931c4290-3d8a-11e9-9c9a-0a580a6c0e78      0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[created-by-prow:true prow.k8s.io/type:periodic prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {[{service {nil nil nil nil nil &SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}] [] [{test gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil} {BUILD_ID 163 nil} {BUILD_NUMBER 163 nil} {JOB_NAME ci-kubernetes-e2e-gke-large-performance-regional nil} {JOB_SPEC {\"type\":\"periodic\",\"job\":\"ci-kubernetes-e2e-gke-large-performance-regional\",\"buildid\":\"163\",\"prowjobid\":\"931c4290-3d8a-11e9-9c9a-0a580a6c0e78\"} nil} {JOB_TYPE periodic nil} {PROW_JOB_ID 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}] Never <nil> <nil>  map[]   0xc012316018  false false false <nil> nil []   nil  [] []  <nil> nil [] <nil>} { []      <nil> [] [] }})",
     "client": "kube",
     "level": "debug",
     "component": "plank"
   },
   "resource": {
     "type": "container",
     "labels": {
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921",
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow"
     }
   },
   "timestamp": "2019-03-03T22:26:54Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T22:27:01.550080143Z"
 },
 {
   "insertId": "1o7ph7cfh465cc",
   "jsonPayload": {
     "msg": "ReplaceProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {931c4290-3d8a-11e9-9c9a-0a580a6c0e78  default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/931c4290-3d8a-11e9-9c9a-0a580a6c0e78 933a4706-3d8a-11e9-898b-42010a80003a 189240028 1 2019-03-03 08:01:39 +0000 UTC <nil> <nil> map[prow.k8s.io/type:periodic created-by-prow:true preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {periodic kubernetes default test-pods ci-kubernetes-e2e-gke-large-performance-regional <nil> [] false   0 false &PodSpec{Volumes:[{service {nil nil nil nil nil SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{ gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}],RestartPolicy:,TerminationGracePeriodSeconds:nil,ActiveDeadlineSeconds:nil,DNSPolicy:,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:nil,ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,} <nil> <nil>} {2019-03-03 08:01:39 +0000 UTC <nil> pending Job triggered. https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-large-performance-regional/162/ 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 162  map[github-reporter:pending]}})",
     "client": "kube",
     "level": "debug",
     "component": "plank"
   },
   "resource": {
     "type": "container",
     "labels": {
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T15:15:55Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T15:15:59.459168535Z"
 },
 {
   "insertId": "1o7ph7cfh465cb",
   "jsonPayload": {
     "client": "kube",
     "level": "debug",
     "component": "plank",
     "msg": "GetProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78)"
   },
   "resource": {
     "type": "container",
     "labels": {
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T15:15:55Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T15:15:59.459168535Z"
 },
 {
   "insertId": "1o7ph7cfh4659z",
   "jsonPayload": {
     "client": "kube",
     "level": "debug",
     "component": "plank",
     "msg": "ReplaceProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {931c4290-3d8a-11e9-9c9a-0a580a6c0e78  default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/931c4290-3d8a-11e9-9c9a-0a580a6c0e78 933a4706-3d8a-11e9-898b-42010a80003a 189117035 1 2019-03-03 08:01:39 +0000 UTC <nil> <nil> map[created-by-prow:true preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional prow.k8s.io/type:periodic] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {periodic kubernetes default test-pods ci-kubernetes-e2e-gke-large-performance-regional <nil> [] false   0 false &PodSpec{Volumes:[{service {nil nil nil nil nil SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{ gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}],RestartPolicy:,TerminationGracePeriodSeconds:nil,ActiveDeadlineSeconds:nil,DNSPolicy:,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:nil,ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,} <nil> <nil>} {2019-03-03 08:01:39 +0000 UTC <nil> pending Job triggered. https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-large-performance-regional/162/ 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 162  map[github-reporter:pending]}})"
   },
   "resource": {
     "type": "container",
     "labels": {
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T15:15:55Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T15:15:59.459168535Z"
 },
 {
   "insertId": "1o7ph7cfh4659y",
   "jsonPayload": {
     "job": "ci-kubernetes-e2e-gke-large-performance-regional",
     "msg": "Pod is missing, starting a new pod",
     "level": "info",
     "name": "931c4290-3d8a-11e9-9c9a-0a580a6c0e78",
     "component": "plank",
     "type": "periodic"
   },
   "resource": {
     "type": "container",
     "labels": {
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T15:15:55Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T15:15:59.459168535Z"
 },
 {
   "insertId": "1o7ph7cfh4659x",
   "jsonPayload": {
     "msg": "CreatePod({{ } {931c4290-3d8a-11e9-9c9a-0a580a6c0e78      0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 preset-k8s-ssh:true created-by-prow:true prow.k8s.io/type:periodic prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {[{service {nil nil nil nil nil &SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}] [] [{test gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil} {BUILD_ID 162 nil} {BUILD_NUMBER 162 nil} {JOB_NAME ci-kubernetes-e2e-gke-large-performance-regional nil} {JOB_SPEC {\"type\":\"periodic\",\"job\":\"ci-kubernetes-e2e-gke-large-performance-regional\",\"buildid\":\"162\",\"prowjobid\":\"931c4290-3d8a-11e9-9c9a-0a580a6c0e78\"} nil} {JOB_TYPE periodic nil} {PROW_JOB_ID 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}] Never <nil> <nil>  map[]   0xc0121e6294  false false false <nil> nil []   nil  [] []  <nil> nil [] <nil>} { []      <nil> [] [] }})",
     "client": "kube",
     "level": "debug",
     "component": "plank"
   },
   "resource": {
     "type": "container",
     "labels": {
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow"
     }
   },
   "timestamp": "2019-03-03T15:15:55Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T15:15:59.459168535Z"
 },
 {
   "insertId": "1vpritkfhzolmx",
   "jsonPayload": {
     "client": "kube",
     "level": "debug",
     "component": "plank",
     "msg": "ReplaceProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {931c4290-3d8a-11e9-9c9a-0a580a6c0e78  default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/931c4290-3d8a-11e9-9c9a-0a580a6c0e78 933a4706-3d8a-11e9-898b-42010a80003a 189117014 1 2019-03-03 08:01:39 +0000 UTC <nil> <nil> map[created-by-prow:true preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional prow.k8s.io/type:periodic] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {periodic kubernetes default test-pods ci-kubernetes-e2e-gke-large-performance-regional <nil> [] false   0 false &PodSpec{Volumes:[{service {nil nil nil nil nil SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{ gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}],RestartPolicy:,TerminationGracePeriodSeconds:nil,ActiveDeadlineSeconds:nil,DNSPolicy:,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:nil,ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,} <nil> <nil>} {2019-03-03 08:01:39 +0000 UTC <nil> pending Job triggered. https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-large-performance-regional/161/ 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 161  map[github-reporter:pending]}})"
   },
   "resource": {
     "type": "container",
     "labels": {
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T08:01:56Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T08:02:01.492391989Z"
 },
 {
   "insertId": "1vpritkfhzolmw",
   "jsonPayload": {
     "component": "plank",
     "msg": "GetProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78)",
     "client": "kube",
     "level": "debug"
   },
   "resource": {
     "type": "container",
     "labels": {
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T08:01:56Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T08:02:01.492391989Z"
 },
 {
   "insertId": "1vpritkfhzollp",
   "jsonPayload": {
     "component": "plank",
     "msg": "ReplaceProwJob(931c4290-3d8a-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {931c4290-3d8a-11e9-9c9a-0a580a6c0e78  default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/931c4290-3d8a-11e9-9c9a-0a580a6c0e78 933a4706-3d8a-11e9-898b-42010a80003a 189116931 1 2019-03-03 08:01:39 +0000 UTC <nil> <nil> map[prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 created-by-prow:true preset-k8s-ssh:true preset-service-account:true prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional prow.k8s.io/type:periodic] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {periodic kubernetes default test-pods ci-kubernetes-e2e-gke-large-performance-regional <nil> [] false   0 false &PodSpec{Volumes:[{service {nil nil nil nil nil SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{ gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}],RestartPolicy:,TerminationGracePeriodSeconds:nil,ActiveDeadlineSeconds:nil,DNSPolicy:,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:nil,ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,} <nil> <nil>} {2019-03-03 08:01:39 +0000 UTC <nil> pending Job triggered. https://prow.k8s.io/view/gcs/kubernetes-jenkins/logs/ci-kubernetes-e2e-gke-large-performance-regional/161/ 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 161  map[]}})",
     "client": "kube",
     "level": "debug"
   },
   "resource": {
     "type": "container",
     "labels": {
       "namespace_id": "default",
       "instance_id": "7000980459144515921",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank"
     }
   },
   "timestamp": "2019-03-03T08:01:55Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T08:02:01.492391989Z"
 },
 {
   "insertId": "1vpritkfhzollo",
   "jsonPayload": {
     "msg": "Transitioning states.",
     "job": "ci-kubernetes-e2e-gke-large-performance-regional",
     "to": "pending",
     "level": "info",
     "name": "931c4290-3d8a-11e9-9c9a-0a580a6c0e78",
     "component": "plank",
     "type": "periodic",
     "from": "triggered"
   },
   "resource": {
     "type": "container",
     "labels": {
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "zone": "us-central1-f",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T08:01:55Z",
   "severity": "ERROR",
   "labels": {
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default",
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T08:02:01.492391989Z"
 },
 {
   "insertId": "1vpritkfhzolln",
   "jsonPayload": {
     "component": "plank",
     "msg": "CreatePod({{ } {931c4290-3d8a-11e9-9c9a-0a580a6c0e78      0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:931c4290-3d8a-11e9-9c9a-0a580a6c0e78 created-by-prow:true prow.k8s.io/type:periodic] map[prow.k8s.io/job:ci-kubernetes-e2e-gke-large-performance-regional] [] nil [] } {[{service {nil nil nil nil nil &SecretVolumeSource{SecretName:service-account,Items:[],DefaultMode:nil,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}} {ssh {nil nil nil nil nil &SecretVolumeSource{SecretName:ssh-key-secret,Items:[],DefaultMode:*256,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}] [] [{test gcr.io/k8s-testimages/kubekins-e2e:v20190301-76bc03340-master [] [--timeout=600 --repo=k8s.io/kubernetes=master --repo=k8s.io/perf-tests=master --root=/go/src --scenario=kubernetes_e2e -- --cluster=gke-regional-cluster --deployment=gke --extract=ci/latest-1.13 --gcp-cloud-sdk=gs://cloud-sdk-testing/ci/staging --gcp-node-image=gci --gcp-project=kubernetes-scale --gcp-region=us-east1 --gke-create-command=container clusters create --quiet --enable-ip-alias --create-subnetwork name=ip-alias-subnet-regional --cluster-ipv4-cidr=/12 --services-ipv4-cidr=/19 --gke-environment=staging --gke-node-locations=us-east1-b --gke-shape={\"default\":{\"Nodes\":1999,\"MachineType\":\"n1-standard-1\"},\"heapster-pool\":{\"Nodes\":1,\"MachineType\":\"n1-standard-8\"}} --provider=gke --test=false --test-cmd=$GOPATH/src/k8s.io/perf-tests/run-e2e.sh --test-cmd-args=cluster-loader2 --test-cmd-args=--nodes=2000 --test-cmd-args=--provider=gke --test-cmd-args=--report-dir=/workspace/_artifacts --test-cmd-args=--testconfig=testing/density/config.yaml --test-cmd-args=--testconfig=testing/load/config.yaml --test-cmd-args=--testoverrides=./testing/density/2000_nodes/override.yaml --test-cmd-name=ClusterLoaderV2 --timeout=570m --use-logexporter]  [] [] [{GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {E2E_GOOGLE_APPLICATION_CREDENTIALS /etc/service-account/service-account.json nil} {USER prow nil} {JENKINS_GCE_SSH_PRIVATE_KEY_FILE /etc/ssh-key-secret/ssh-private nil} {JENKINS_GCE_SSH_PUBLIC_KEY_FILE /etc/ssh-key-secret/ssh-public nil} {BUILD_ID 161 nil} {BUILD_NUMBER 161 nil} {JOB_NAME ci-kubernetes-e2e-gke-large-performance-regional nil} {JOB_SPEC {\"type\":\"periodic\",\"job\":\"ci-kubernetes-e2e-gke-large-performance-regional\",\"buildid\":\"161\",\"prowjobid\":\"931c4290-3d8a-11e9-9c9a-0a580a6c0e78\"} nil} {JOB_TYPE periodic nil} {PROW_JOB_ID 931c4290-3d8a-11e9-9c9a-0a580a6c0e78 nil}] {map[] map[cpu:{{6 0} {<nil>} 6 DecimalSI} memory:{{17179869184 0} {<nil>}  BinarySI}]} [{service true /etc/service-account  <nil>} {ssh true /etc/ssh-key-secret  <nil>}] [] nil nil nil    nil false false false}] Never <nil> <nil>  map[]   0xc006506a04  false false false <nil> nil []   nil  [] []  <nil> nil [] <nil>} { []      <nil> [] [] }})",
     "client": "kube",
     "level": "debug"
   },
   "resource": {
     "type": "container",
     "labels": {
       "zone": "us-central1-f",
       "pod_id": "plank-9f6cb7fbb-4jdf2",
       "project_id": "k8s-prow",
       "cluster_name": "prow",
       "container_name": "plank",
       "namespace_id": "default",
       "instance_id": "7000980459144515921"
     }
   },
   "timestamp": "2019-03-03T08:01:55Z",
   "severity": "ERROR",
   "labels": {
     "compute.googleapis.com/resource_name": "fluentd-gcp-v3.2.0-56c8g",
     "container.googleapis.com/pod_name": "plank-9f6cb7fbb-4jdf2",
     "container.googleapis.com/stream": "stderr",
     "container.googleapis.com/namespace_name": "default"
   },
   "logName": "projects/k8s-prow/logs/plank",
   "receiveTimestamp": "2019-03-03T08:02:01.492391989Z"
 }
]

@stevekuznetsov
Copy link
Contributor

Wait, so were there actually multiple versions of the test running at once?

@mm4tt
Copy link
Contributor Author

mm4tt commented Mar 4, 2019

Matt Matejczyk FYI the best place to escalate something like this is in #testing-ops on Slack, pinging @test-infra-oncall. I've shot a message over here

Thanks, @stevekuznetsov. Will keep that in mind for the future.

There are other jobs behaving similarly to this one, i.e. they are scheduled and run when they shouldn't be run.

could you list them?

@BenTheElder, other examples

name: ci-kubernetes-e2e-gce-scale-performance
config
rbkkcuqztpa

Job should be run once Mon-Fri, but recently there are days when it's run twice or thrice:
d0dofwbipr5

name: ci-kubernetes-e2e-gke-large-performance
config
weyhn28fg2c

Job is supposed to be run once every Sunday, but last yesterday was launched twice:
k4njvwksfjk

There are probably a few more.

Were you able to figure out what is going on?

@krzyzacy
Copy link
Member

krzyzacy commented Mar 4, 2019

/cc

@krzyzacy
Copy link
Member

krzyzacy commented Mar 4, 2019

  component:  "plank"   
  job:  "ci-kubernetes-e2e-gce-scale-performance"   
  level:  "info"   
  msg:  "Pod is missing, starting a new pod"   

We probably hit the OOMKilled again?

@krzyzacy
Copy link
Member

krzyzacy commented Mar 4, 2019

E  I0304 18:02:18.055] Call:  gsutil -q -h Content-Type:application/json -h x-goog-if-generation-match:1551706405925783 cp /tmp/gsutil_h1mRnW gs://kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-scale-performance/jobResultsCache.json 
E  I0304 18:02:19.881] process 693067 exited with code 0 after 0.0m 
E  I0304 18:02:19.884] Call:  gsutil -q -h Content-Type:application/json cp /tmp/gsutil_7WyV5x gs://kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-scale-performance/324/finished.json 
E  I0304 18:02:21.640] process 693245 exited with code 0 after 0.0m 
E  I0304 18:02:21.641] Call:  gsutil -q -h Content-Type:text/plain -h 'Cache-Control:private, max-age=0, no-transform' cp /tmp/gsutil_tpumdl gs://kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-scale-performance/latest-build.txt 
E  I0304 18:02:23.301] process 693423 exited with code 0 after 0.0m 
E  I0304 18:02:23.302] Call:  gsutil -q cp -Z /workspace/build-log.txt gs://kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-scale-performance/324/build-log.txt 
E  I0304 18:02:25.317] process 693601 exited with code 0 after 0.0m 
E  + EXIT_VALUE=1 
E  + set +o xtrace 
E  Cloning into 'test-infra'... 
E  Activated service account credentials for: [[email protected]] 
E  fatal: Not a git repository (or any of the parent directories): .git 

@stevekuznetsov seems the pod finished and exited properly? Seems a bug in plank?

@krzyzacy
Copy link
Member

krzyzacy commented Mar 4, 2019

Horologium triggered the job properly afaik

also - no associated logs in sinker (so - who deleted the pod?)

@stevekuznetsov
Copy link
Contributor

@krzyzacy what was the behavior? Plank will create a Pod if one does not exist and the ProwJob is not marked in some completed state, can you try to determine via logs how the pod exited and what the state of the prowjob was at the time?

@krzyzacy
Copy link
Member

krzyzacy commented Mar 4, 2019

The pod exited with 1 (with E + EXIT_VALUE=1 ) I believe...

I think the problem occurs after #11477? (Feb.26 according to @mm4tt 's screenshot)

And I think the prowjob was still in pending state as I don't see any other state transition log:

2019-03-04 00:01:53.000 PST
{"msg":"CreatePod({{ } {bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[prow.k8s.io/id:bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 created-by-prow:true prow.k8s.io/type:periodic prow.k8s.io/job:ci-kubernetes-e2e-gce-scale-performance preset-e2e-scalability-common:t…
2019-03-04 00:01:53.000 PST
{"component":"plank","type":"periodic","from":"triggered","msg":"Transitioning states.","to":"pending","job":"ci-kubernetes-e2e-gce-scale-performance","level":"info","name":"bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78"}
2019-03-04 00:01:53.000 PST
{"msg":"ReplaceProwJob(bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 bc8b31ae-3e53-11e9-898b-42010a80003a 189526855 1 2019-03-04 08:01:37 +0000 UTC <…
2019-03-04 00:01:54.000 PST
{"component":"plank","msg":"GetProwJob(bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78)","client":"kube","level":"debug"}
2019-03-04 00:01:54.000 PST
{"component":"plank","msg":"ReplaceProwJob(bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 bc8b31ae-3e53-11e9-898b-42010a80003a 189526926 1 2019-03-04 …
2019-03-04 05:33:53.000 PST
{"msg":"CreatePod({{ } {bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[prow.k8s.io/type:periodic prow.k8s.io/job:ci-kubernetes-e2e-gce-scale-performance preset-e2e-scalability-common:true preset-k8s-ssh:true preset-service-account:true prow.k8s.io/id:bc8b0b…
2019-03-04 05:33:53.000 PST
{"msg":"Pod is missing, starting a new pod","job":"ci-kubernetes-e2e-gce-scale-performance","level":"info","name":"bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78","type":"periodic","component":"plank"}
2019-03-04 05:33:53.000 PST
{"component":"plank","msg":"ReplaceProwJob(bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 bc8b31ae-3e53-11e9-898b-42010a80003a 189526934 1 2019-03-04 …
2019-03-04 05:33:53.000 PST
{"component":"plank","msg":"GetProwJob(bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78)","client":"kube","level":"debug"}
 
2019-03-04 05:33:53.000 PST
{"component":"plank","msg":"ReplaceProwJob(bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 bc8b31ae-3e53-11e9-898b-42010a80003a 189621384 1 2019-03-04 …
2019-03-04 10:02:54.000 PST
{"client":"kube","level":"debug","component":"plank","msg":"CreatePod({{ } {bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[prow.k8s.io/id:bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 created-by-prow:true prow.k8s.io/type:periodic prow.k8s.io/job:ci-kubernetes-e2e-g…
2019-03-04 10:02:54.000 PST
{"msg":"Pod is missing, starting a new pod","job":"ci-kubernetes-e2e-gce-scale-performance","level":"info","name":"bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78","component":"plank","type":"periodic"}
2019-03-04 10:02:54.000 PST
{"component":"plank","msg":"ReplaceProwJob(bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 bc8b31ae-3e53-11e9-898b-42010a80003a 189621384 1 2019-03-04 …
2019-03-04 10:02:55.000 PST
{"msg":"GetProwJob(bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78)","client":"kube","level":"debug","component":"plank"}
2019-03-04 10:02:55.000 PST
{"component":"plank","msg":"ReplaceProwJob(bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78, {{ProwJob prow.k8s.io/v1} {bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 default /apis/prow.k8s.io/v1/namespaces/default/prowjobs/bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78 bc8b31ae-3e53-11e9-898b-42010a80003a 189698498 1 2019-03-04 …

@krzyzacy
Copy link
Member

krzyzacy commented Mar 4, 2019

Edit: Just paste the full node log here:

Mar 04 18:02:25 gke-prow-containerd-pool-99179761-9sg5 containerd[1141]: time="2019-03-04T18:02:25Z" level=info msg="Finish piping stderr of container "edeb1523a687b7e5a80ca831b9760a1a6328be767b35a3197d6919752681fc2b""
Mar 04 18:02:25 gke-prow-containerd-pool-99179761-9sg5 containerd[1141]: time="2019-03-04T18:02:25Z" level=info msg="Finish piping stdout of container "edeb1523a687b7e5a80ca831b9760a1a6328be767b35a3197d6919752681fc2b""
Mar 04 18:02:25 gke-prow-containerd-pool-99179761-9sg5 containerd[1141]: time="2019-03-04T18:02:25Z" level=error msg="collecting metrics for edeb1523a687b7e5a80ca831b9760a1a6328be767b35a3197d6919752681fc2b" error="cgroups: cgroup deleted"
Mar 04 18:02:25 gke-prow-containerd-pool-99179761-9sg5 containerd[1141]: time="2019-03-04T18:02:25Z" level=info msg="shim reaped" id=edeb1523a687b7e5a80ca831b9760a1a6328be767b35a3197d6919752681fc2b
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:26.074868    1260 kubelet.go:1883] SyncLoop (PLEG): "bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78_test-pods(2763ab6d-3e82-11e9-989e-42010a800133)", event: &pleg.PodLifecycleEvent{ID:"2763ab6d-3e82-11e9-989e-42010a800133", Type:"ContainerDied", Data:"edeb1523a687b7e5a80ca831b9760a1a6328be767b35a3197d6919752681fc2b"}
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 containerd[1141]: time="2019-03-04T18:02:26Z" level=info msg="StopPodSandbox for "b1fee6999e358a3e062e04f40ff5be40ac9d89b96146ece5e0ff8c541b485a4e""
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 containerd[1141]: time="2019-03-04T18:02:26Z" level=info msg="Container to stop "edeb1523a687b7e5a80ca831b9760a1a6328be767b35a3197d6919752681fc2b" is not running, current state "CONTAINER_EXITED""
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 systemd-networkd[350]: veth57837a20: Lost carrier
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 systemd-timesyncd[314]: Network configuration changed, trying to establish connection.
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 systemd-timesyncd[314]: Synchronized to time server 169.254.169.254:123 (169.254.169.254).
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 systemd-timesyncd[314]: Network configuration changed, trying to establish connection.
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 systemd-timesyncd[314]: Synchronized to time server 169.254.169.254:123 (169.254.169.254).
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 containerd[1141]: time="2019-03-04T18:02:26Z" level=info msg="TearDown network for sandbox "b1fee6999e358a3e062e04f40ff5be40ac9d89b96146ece5e0ff8c541b485a4e" successfully"
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:26.228781    1260 reconciler.go:181] operationExecutor.UnmountVolume started for volume "ssh" (UniqueName: "kubernetes.io/secret/2763ab6d-3e82-11e9-989e-42010a800133-ssh") pod "2763ab6d-3e82-11e9-989e-42010a800133" (UID: "2763ab6d-3e82-11e9-989e-42010a800133")
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:26.228850    1260 reconciler.go:181] operationExecutor.UnmountVolume started for volume "service" (UniqueName: "kubernetes.io/secret/2763ab6d-3e82-11e9-989e-42010a800133-service") pod "2763ab6d-3e82-11e9-989e-42010a800133" (UID: "2763ab6d-3e82-11e9-989e-42010a800133")
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:26.246068    1260 operation_generator.go:688] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2763ab6d-3e82-11e9-989e-42010a800133-ssh" (OuterVolumeSpecName: "ssh") pod "2763ab6d-3e82-11e9-989e-42010a800133" (UID: "2763ab6d-3e82-11e9-989e-42010a800133"). InnerVolumeSpecName "ssh". PluginName "kubernetes.io/secret", VolumeGidValue ""
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:26.247230    1260 operation_generator.go:688] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2763ab6d-3e82-11e9-989e-42010a800133-service" (OuterVolumeSpecName: "service") pod "2763ab6d-3e82-11e9-989e-42010a800133" (UID: "2763ab6d-3e82-11e9-989e-42010a800133"). InnerVolumeSpecName "service". PluginName "kubernetes.io/secret", VolumeGidValue ""
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:26.329170    1260 reconciler.go:301] Volume detached for volume "service" (UniqueName: "kubernetes.io/secret/2763ab6d-3e82-11e9-989e-42010a800133-service") on node "gke-prow-containerd-pool-99179761-9sg5" DevicePath ""
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:26.329219    1260 reconciler.go:301] Volume detached for volume "ssh" (UniqueName: "kubernetes.io/secret/2763ab6d-3e82-11e9-989e-42010a800133-ssh") on node "gke-prow-containerd-pool-99179761-9sg5" DevicePath ""
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 containerd[1141]: time="2019-03-04T18:02:26Z" level=info msg="shim reaped" id=b1fee6999e358a3e062e04f40ff5be40ac9d89b96146ece5e0ff8c541b485a4e
Mar 04 18:02:26 gke-prow-containerd-pool-99179761-9sg5 containerd[1141]: time="2019-03-04T18:02:26Z" level=info msg="StopPodSandbox for "b1fee6999e358a3e062e04f40ff5be40ac9d89b96146ece5e0ff8c541b485a4e" returns successfully"
Mar 04 18:02:27 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:27.075432    1260 kubelet.go:1883] SyncLoop (PLEG): "bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78_test-pods(2763ab6d-3e82-11e9-989e-42010a800133)", event: &pleg.PodLifecycleEvent{ID:"2763ab6d-3e82-11e9-989e-42010a800133", Type:"ContainerDied", Data:"b1fee6999e358a3e062e04f40ff5be40ac9d89b96146ece5e0ff8c541b485a4e"}
Mar 04 18:02:27 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: W0304 18:02:27.075574    1260 pod_container_deletor.go:75] Container "b1fee6999e358a3e062e04f40ff5be40ac9d89b96146ece5e0ff8c541b485a4e" not found in pod's containers
Mar 04 18:02:33 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:33.622347    1260 kubelet.go:1854] SyncLoop (DELETE, "api"): "bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78_test-pods(2763ab6d-3e82-11e9-989e-42010a800133)"
Mar 04 18:02:33 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:33.623922    1260 kubelet.go:1848] SyncLoop (REMOVE, "api"): "bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78_test-pods(2763ab6d-3e82-11e9-989e-42010a800133)"
Mar 04 18:02:33 gke-prow-containerd-pool-99179761-9sg5 kubelet[1260]: I0304 18:02:33.624053    1260 kubelet.go:2042] Failed to delete pod "bc8b0b06-3e53-11e9-9c9a-0a580a6c0e78_test-pods(2763ab6d-3e82-11e9-989e-42010a800133)", err: pod not found

@stevekuznetsov
Copy link
Contributor

If the pod exited, are you seeing sinker clean it up? If not, do you have audit logging on? Do you know what deleted the Pod?

@krzyzacy
Copy link
Member

krzyzacy commented Mar 4, 2019

I don't see the clean up from sinker. How do I check the audit log?

@stevekuznetsov
Copy link
Contributor

@krzyzacy
Copy link
Member

@wojtek-t by "separate build cluster" we mean a cluster that we schedule all scalaibility job pods there. They will still share the same prow controllers, but you will have control over your secrets/pods/nodes.

(aka k8s prow can schedule pods onto your cluster)

@cjwagner
Copy link
Member

Details: https://github.com/kubernetes/test-infra/blob/master/prow/getting_started_deploy.md#run-test-pods-in-different-clusters

@wojtek-t
Copy link
Member

So I think I'm not following why this is supposed to help. If we would be sharing the same prow controllers, then the limits that we are affected by won't change, right [we're being hit by the concurrency limit at the prow level, right?]

@krzyzacy
Copy link
Member

@wojtek-t nope - the podgc kicks in in the build cluster where all the workloads are running, not in the service cluster.

@wojtek-t
Copy link
Member

Wait - I'm lost then. You're talking about Kubernetes podgc? This should never remove pods in state Running, which is happening in this case.

@krzyzacy
Copy link
Member

krzyzacy commented May 17, 2019

@wojtek-t Prow has a logic that watches completed pod, and update prowjob status as passed/failed/etc. Right when a pod is finished, it's got deleted by K8s podgc, and plank does not using informer cache yet so it will miss the event, saw pod associated with the prowjob is missing, and restart the pod.

(aka we tries to manage pod lifecycle by prow, not K8s podgc, and now they are fighting.)

@wojtek-t
Copy link
Member

@krzyzacy - I'm not entirely sure it's our problem then, because unless I'm not missing anything we've seen cases where the test was restarted in the middle of the run, not after it was finished...

But anyway - I'm fine with using a separate cluster, because it will mitigate at least some of the issues...

@krzyzacy
Copy link
Member

@wojtek-t I think it shows up as 'R' in testgrid because the log failed to be uploaded, as testgrid is basically reading gcs. I can also take a closer look.

@wojtek-t
Copy link
Member

@wojtek-t I think it shows up as 'R' in testgrid because the log failed to be uploaded, as testgrid is basically reading gcs. I can also take a closer look.

This I know (and yes that happens). But we've seen cases that it is restarted in the meantime (though it's hard to tell why it was restarted and what happened in that first round) and the restarted attempt finishes on time and is reported correctly. Though I don't know if it was happening recently too..

@krzyzacy
Copy link
Member

@wojtek-t Or we might be hitting disk pressure again, I think we never solved that with the logexporter. Use dedicated nodes to run scalability jobs and give them more resources should help in general.

I know you want to own your cluster, maybe hand us a cluster credential and we can help with carrying over the existing secrets?

@wojtek-t
Copy link
Member

I know you want to own your cluster, maybe hand us a cluster credential and we can help with carrying over the existing secrets?

Sounds reasonable. Though it my be after kubecon, unless @krzysied or @mborsz would like to help with that.

@oxddr
Copy link
Contributor

oxddr commented Jun 3, 2019

This has hit us again. I've set up us a dedicated cluster in which we'd like to run scalability jobs.

@krzyzacy I'll follow up with you to pass the credentials.

@oxddr
Copy link
Contributor

oxddr commented Jun 6, 2019

After switching to a dedicated build cluster scalability tests still suffers from the same problem (or we see the same symptoms).
7jmUc0L9vhV

@oxddr
Copy link
Contributor

oxddr commented Jun 6, 2019

FYI: I've changed machines in the dedicated clusters to 16 cores, as I've seen there was ~100% CPU usage. We need to wait yet another day to check whether it helped.

@krzyzacy
Copy link
Member

krzyzacy commented Jun 6, 2019

this smells like a different pattern, the job suppose to run daily?

@stevekuznetsov
Copy link
Contributor

Since you guys have your own cluster now please use your credentials to investigate the test pods and pod events to see what is happening.

@krzyzacy
Copy link
Member

krzyzacy commented Jun 6, 2019

this smells the pod was evicted

2019-06-05 18:52:56.000 PDT
The node was low on resource: ephemeral-storage. Container test was using 223620144Ki, which exceeds its request of 0.

The last few line of logs are like

E  W0606 01:51:36.602] scp: /var/log/fluentd.log*: No such file or directory
 
E  W0606 01:51:36.603] scp: /var/log/node-problem-detector.log*: No such file or directory
 
E  W0606 01:51:36.603] scp: /var/log/kubelet.cov*: No such file or directory
 
E  W0606 01:51:36.603] scp: /var/log/startupscript.log*: No such file or directory
 
E  W0606 01:51:36.607] ERROR: (gcloud.compute.scp) [/usr/bin/scp] exited with return code [1].

which indicates log-exporter was not working properly?

And kubelet log:

MESSAGE:  "I0606 01:53:06.968036    1592 eviction_manager.go:191] eviction manager: pods 26636eb1-8768-11e9-83de-3eff4e481367_test-pods(07c2a030-87cf-11e9-b0b1-42010a8000d8) evicted, waiting for pod to be cleaned up"   

@BenTheElder
Copy link
Member

@oxddr 223620144Ki -> 229 GB, you're using more storage than the node can fit on disk ... sounds like we're pulling down logs to the job again, instead of up to GCS. That's just not going to work well. You could work around it by making nodes with very large disks and adding proper resource specs to the pods

@oxddr
Copy link
Contributor

oxddr commented Jun 6, 2019

Thank you all for help and tips. Hopefully next time, I'll be able to debug it myself.

I've increase disk size with the 16 core nodes added earlier today, but I am going to bump it even further to make sure our tests passes. I'll have a look tomorrow into why we are using so much disk space.

@mm4tt
Copy link
Contributor Author

mm4tt commented Jun 7, 2019

It's strange that we've started using more storage, given that the only change to our job config was adding

  cluster: scalability

Maybe it's some kind of a permission issue, where due to misconfigured/missing secrets we cannot write to GCS and fallback to dumping the logs locally?

@oxddr
Copy link
Contributor

oxddr commented Jun 7, 2019

This is clearly a separate issue. I've opened #12940, so this can be tracked there.

@krzyzacy
Copy link
Member

(seems not happening anymore? feel free to reopen)

@mm4tt
Copy link
Contributor Author

mm4tt commented Sep 30, 2019

Yes, since we switched to our own, dedicated build cluster it stopped happening. There are some "cleanups" we need to do to finalize this, but they're tracked in separate issues. We can close this one.

Thanks, Sen!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/bug Categorizes issue or PR as related to a bug. priority/important-soon Must be staffed and worked on either currently, or very soon, ideally in time for the next release. sig/scalability Categorizes an issue or PR as relevant to SIG Scalability. sig/testing Categorizes an issue or PR as relevant to SIG Testing.
Projects
None yet
Development

No branches or pull requests