diff --git a/deploy/addons/aliyun_mirror.json b/deploy/addons/aliyun_mirror.json index 4ce5afae0535..02b471a2395c 100644 --- a/deploy/addons/aliyun_mirror.json +++ b/deploy/addons/aliyun_mirror.json @@ -4,7 +4,7 @@ "gcr.io/k8s-minikube/auto-pause-hook": "registry.cn-hangzhou.aliyuncs.com/google_containers/auto-pause-hook", "quay.io/operator-framework/olm": "registry.cn-hangzhou.aliyuncs.com/google_containers/olm", "quay.io/operator-framework/upstream-community-operators": "registry.cn-hangzhou.aliyuncs.com/google_containers/upstream-community-operators", - "k8s.gcr.io/kube-registry-proxy": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-registry-proxy", + "registry.k8s.io/kube-registry-proxy": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-registry-proxy", "docker.io/upmcenterprises/registry-creds": "registry.cn-hangzhou.aliyuncs.com/google_containers/registry-creds", "quay.io/rhdevelopers/core-dns-patcher": "registry.cn-hangzhou.aliyuncs.com/google_containers/core-dns-patcher", "docker.io/nvidia/k8s-device-plugin": "registry.cn-hangzhou.aliyuncs.com/google_containers/k8s-device-plugin", @@ -13,75 +13,75 @@ "quay.io/datawire/ambassador-operator": "registry.cn-hangzhou.aliyuncs.com/google_containers/ambassador-operator", "docker.io/jettech/kube-webhook-certgen": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-webhook-certgen", "gcr.io/k8s-minikube/gcp-auth-webhook": "registry.cn-hangzhou.aliyuncs.com/google_containers/gcp-auth-webhook", - "k8s.gcr.io/sig-storage/snapshot-controller": "registry.cn-hangzhou.aliyuncs.com/google_containers/snapshot-controller", - "k8s.gcr.io/sig-storage/csi-attacher": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-attacher", - "k8s.gcr.io/sig-storage/csi-external-health-monitor-agent": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-external-health-monitor-agent", - "k8s.gcr.io/sig-storage/csi-external-health-monitor-controller": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-external-health-monitor-controller", - "k8s.gcr.io/sig-storage/csi-node-driver-registrar": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-node-driver-registrar", - "k8s.gcr.io/sig-storage/hostpathplugin": "registry.cn-hangzhou.aliyuncs.com/google_containers/hostpathplugin", - "k8s.gcr.io/sig-storage/livenessprobe": "registry.cn-hangzhou.aliyuncs.com/google_containers/livenessprobe", - "k8s.gcr.io/sig-storage/csi-resizer": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-resizer", - "k8s.gcr.io/sig-storage/csi-snapshotter": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-snapshotter", - "k8s.gcr.io/sig-storage/csi-provisioner": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-provisioner", + "registry.k8s.io/sig-storage/snapshot-controller": "registry.cn-hangzhou.aliyuncs.com/google_containers/snapshot-controller", + "registry.k8s.io/sig-storage/csi-attacher": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-attacher", + "registry.k8s.io/sig-storage/csi-external-health-monitor-agent": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-external-health-monitor-agent", + "registry.k8s.io/sig-storage/csi-external-health-monitor-controller": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-external-health-monitor-controller", + "registry.k8s.io/sig-storage/csi-node-driver-registrar": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-node-driver-registrar", + "registry.k8s.io/sig-storage/hostpathplugin": "registry.cn-hangzhou.aliyuncs.com/google_containers/hostpathplugin", + "registry.k8s.io/sig-storage/livenessprobe": "registry.cn-hangzhou.aliyuncs.com/google_containers/livenessprobe", + "registry.k8s.io/sig-storage/csi-resizer": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-resizer", + "registry.k8s.io/sig-storage/csi-snapshotter": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-snapshotter", + "registry.k8s.io/sig-storage/csi-provisioner": "registry.cn-hangzhou.aliyuncs.com/google_containers/csi-provisioner", "docker.io/registry": "registry.cn-hangzhou.aliyuncs.com/google_containers/registry", "docker.io/gluster/gluster-centos": "registry.cn-hangzhou.aliyuncs.com/google_containers/glusterfs-server", "docker.io/heketi/heketi": "registry.cn-hangzhou.aliyuncs.com/google_containers/heketi", "docker.io/coredns/coredns": "registry.cn-hangzhou.aliyuncs.com/google_containers/coredns", "docker.io/kindest/kindnetd": "registry.cn-hangzhou.aliyuncs.com/google_containers/kindnetd", - "k8s.gcr.io/ingress-nginx/controller": "registry.cn-hangzhou.aliyuncs.com/google_containers/nginx-ingress-controller", + "registry.k8s.io/ingress-nginx/controller": "registry.cn-hangzhou.aliyuncs.com/google_containers/nginx-ingress-controller", "gcr.io/cloud-builders/gcs-fetcher": "registry.cn-hangzhou.aliyuncs.com/cloud-builders/gcs-fetcher", "gcr.io/google-samples/freshpod": "registry.cn-hangzhou.aliyuncs.com/google_containers/freshpod", "gcr.io/k8s-minikube/gvisor-addon": "registry.cn-hangzhou.aliyuncs.com/google_containers/gvisor-addon", "gcr.io/k8s-minikube/kicbase": "registry.cn-hangzhou.aliyuncs.com/google_containers/kicbase", "gcr.io/k8s-minikube/storage-provisioner": "registry.cn-hangzhou.aliyuncs.com/google_containers/storage-provisioner", "gcr.io/kubernetes-helm/tiller": "registry.cn-hangzhou.aliyuncs.com/google_containers/tiller", - "k8s.gcr.io/addon-resizer": "registry.cn-hangzhou.aliyuncs.com/google_containers/addon-resizer", - "k8s.gcr.io/busybox": "registry.cn-hangzhou.aliyuncs.com/google_containers/busybox", - "k8s.gcr.io/cluster-autoscaler": "registry.cn-hangzhou.aliyuncs.com/google_containers/cluster-autoscaler", - "k8s.gcr.io/coredns/coredns": "registry.cn-hangzhou.aliyuncs.com/google_containers/coredns", - "k8s.gcr.io/defaultbackend": "registry.cn-hangzhou.aliyuncs.com/google_containers/defaultbackend", - "k8s.gcr.io/echoserver": "registry.cn-hangzhou.aliyuncs.com/google_containers/echoserver", - "k8s.gcr.io/elasticsearch": "registry.cn-hangzhou.aliyuncs.com/google_containers/elasticsearch", - "k8s.gcr.io/etcd": "registry.cn-hangzhou.aliyuncs.com/google_containers/etcd", - "k8s.gcr.io/etcd-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/etcd-amd64", - "k8s.gcr.io/exechealthz-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/exechealthz-amd64", - "k8s.gcr.io/flannel-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/flannel-amd64", - "k8s.gcr.io/fluentd-elasticsearch": "registry.cn-hangzhou.aliyuncs.com/google_containers/fluentd-elasticsearch", - "k8s.gcr.io/heapster": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster", - "k8s.gcr.io/heapster_grafana": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster_grafana", - "k8s.gcr.io/heapster_influxdb": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster_influxdb", - "k8s.gcr.io/heapster-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster-amd64", - "k8s.gcr.io/heapster-grafana-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster-grafana-amd64", - "k8s.gcr.io/heapster-influxdb-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster-influxdb-amd64", - "k8s.gcr.io/k8s-dns-dnsmasq-nanny-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/k8s-dns-dnsmasq-nanny-amd64", - "k8s.gcr.io/k8s-dns-kube-dns-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/k8s-dns-kube-dns-amd64", - "k8s.gcr.io/k8s-dns-node-cache": "registry.cn-hangzhou.aliyuncs.com/google_containers/k8s-dns-node-cache", - "k8s.gcr.io/k8s-dns-sidecar-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/k8s-dns-sidecar-amd64", - "k8s.gcr.io/kube-addon-manager": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-addon-manager", - "k8s.gcr.io/kube-addon-manager-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-addon-manager-amd64", - "k8s.gcr.io/kube-apiserver": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-apiserver", - "k8s.gcr.io/kube-apiserver-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-apiserver-amd64", - "k8s.gcr.io/kube-controller-manager": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-controller-manager", - "k8s.gcr.io/kube-controller-manager-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-controller-manager-amd64", - "k8s.gcr.io/kube-cross": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-cross", - "k8s.gcr.io/kube-dnsmasq-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-dnsmasq-amd64", - "k8s.gcr.io/kube-proxy": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-proxy", - "k8s.gcr.io/kube-proxy-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-proxy-amd64", - "k8s.gcr.io/kube-scheduler": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-scheduler", - "k8s.gcr.io/kube-scheduler-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-scheduler-amd64", - "k8s.gcr.io/kube-state-metrics": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-state-metrics", - "k8s.gcr.io/kubedns-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kubedns-amd64", - "k8s.gcr.io/kubernetes-dashboard-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kubernetes-dashboard-amd64", - "k8s.gcr.io/metrics-server-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/metrics-server-amd64", - "k8s.gcr.io/minikube-nvidia-driver-installer": "registry.cn-hangzhou.aliyuncs.com/google_containers/minikube-nvidia-driver-installer", - "k8s.gcr.io/mongodb-install": "registry.cn-hangzhou.aliyuncs.com/google_containers/mongodb-install", - "k8s.gcr.io/nginx-slim": "registry.cn-hangzhou.aliyuncs.com/google_containers/nginx-slim", - "k8s.gcr.io/nvidia-gpu-device-plugin": "registry.cn-hangzhou.aliyuncs.com/google_containers/nvidia-gpu-device-plugin", - "k8s.gcr.io/pause": "registry.cn-hangzhou.aliyuncs.com/google_containers/pause", - "k8s.gcr.io/pause-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/pause-amd64", - "k8s.gcr.io/spark": "registry.cn-hangzhou.aliyuncs.com/google_containers/spark", - "k8s.gcr.io/spartakus-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/spartakus-amd64", - "k8s.gcr.io/zeppelin": "registry.cn-hangzhou.aliyuncs.com/google_containers/zeppelin", + "registry.k8s.io/addon-resizer": "registry.cn-hangzhou.aliyuncs.com/google_containers/addon-resizer", + "registry.k8s.io/busybox": "registry.cn-hangzhou.aliyuncs.com/google_containers/busybox", + "registry.k8s.io/cluster-autoscaler": "registry.cn-hangzhou.aliyuncs.com/google_containers/cluster-autoscaler", + "registry.k8s.io/coredns/coredns": "registry.cn-hangzhou.aliyuncs.com/google_containers/coredns", + "registry.k8s.io/defaultbackend": "registry.cn-hangzhou.aliyuncs.com/google_containers/defaultbackend", + "registry.k8s.io/echoserver": "registry.cn-hangzhou.aliyuncs.com/google_containers/echoserver", + "registry.k8s.io/elasticsearch": "registry.cn-hangzhou.aliyuncs.com/google_containers/elasticsearch", + "registry.k8s.io/etcd": "registry.cn-hangzhou.aliyuncs.com/google_containers/etcd", + "registry.k8s.io/etcd-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/etcd-amd64", + "registry.k8s.io/exechealthz-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/exechealthz-amd64", + "registry.k8s.io/flannel-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/flannel-amd64", + "registry.k8s.io/fluentd-elasticsearch": "registry.cn-hangzhou.aliyuncs.com/google_containers/fluentd-elasticsearch", + "registry.k8s.io/heapster": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster", + "registry.k8s.io/heapster_grafana": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster_grafana", + "registry.k8s.io/heapster_influxdb": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster_influxdb", + "registry.k8s.io/heapster-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster-amd64", + "registry.k8s.io/heapster-grafana-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster-grafana-amd64", + "registry.k8s.io/heapster-influxdb-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/heapster-influxdb-amd64", + "registry.k8s.io/k8s-dns-dnsmasq-nanny-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/k8s-dns-dnsmasq-nanny-amd64", + "registry.k8s.io/k8s-dns-kube-dns-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/k8s-dns-kube-dns-amd64", + "registry.k8s.io/k8s-dns-node-cache": "registry.cn-hangzhou.aliyuncs.com/google_containers/k8s-dns-node-cache", + "registry.k8s.io/k8s-dns-sidecar-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/k8s-dns-sidecar-amd64", + "registry.k8s.io/kube-addon-manager": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-addon-manager", + "registry.k8s.io/kube-addon-manager-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-addon-manager-amd64", + "registry.k8s.io/kube-apiserver": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-apiserver", + "registry.k8s.io/kube-apiserver-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-apiserver-amd64", + "registry.k8s.io/kube-controller-manager": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-controller-manager", + "registry.k8s.io/kube-controller-manager-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-controller-manager-amd64", + "registry.k8s.io/kube-cross": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-cross", + "registry.k8s.io/kube-dnsmasq-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-dnsmasq-amd64", + "registry.k8s.io/kube-proxy": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-proxy", + "registry.k8s.io/kube-proxy-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-proxy-amd64", + "registry.k8s.io/kube-scheduler": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-scheduler", + "registry.k8s.io/kube-scheduler-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-scheduler-amd64", + "registry.k8s.io/kube-state-metrics": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-state-metrics", + "registry.k8s.io/kubedns-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kubedns-amd64", + "registry.k8s.io/kubernetes-dashboard-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/kubernetes-dashboard-amd64", + "registry.k8s.io/metrics-server-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/metrics-server-amd64", + "registry.k8s.io/minikube-nvidia-driver-installer": "registry.cn-hangzhou.aliyuncs.com/google_containers/minikube-nvidia-driver-installer", + "registry.k8s.io/mongodb-install": "registry.cn-hangzhou.aliyuncs.com/google_containers/mongodb-install", + "registry.k8s.io/nginx-slim": "registry.cn-hangzhou.aliyuncs.com/google_containers/nginx-slim", + "registry.k8s.io/nvidia-gpu-device-plugin": "registry.cn-hangzhou.aliyuncs.com/google_containers/nvidia-gpu-device-plugin", + "registry.k8s.io/pause": "registry.cn-hangzhou.aliyuncs.com/google_containers/pause", + "registry.k8s.io/pause-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/pause-amd64", + "registry.k8s.io/spark": "registry.cn-hangzhou.aliyuncs.com/google_containers/spark", + "registry.k8s.io/spartakus-amd64": "registry.cn-hangzhou.aliyuncs.com/google_containers/spartakus-amd64", + "registry.k8s.io/zeppelin": "registry.cn-hangzhou.aliyuncs.com/google_containers/zeppelin", "quay.io/coreos/configmap-reload": "registry.cn-hangzhou.aliyuncs.com/coreos_containers/configmap-reload", "quay.io/coreos/grafana-watcher": "registry.cn-hangzhou.aliyuncs.com/coreos_containers/grafana-watcher", "quay.io/coreos/hyperkube": "registry.cn-hangzhou.aliyuncs.com/coreos_containers/hyperkube", @@ -94,9 +94,9 @@ "quay.io/kubernetes-service-catalog/service-catalog": "registry.cn-hangzhou.aliyuncs.com/kubernetes-service-catalog/service-catalog", "quay.io/prometheus/alertmanager": "registry.cn-hangzhou.aliyuncs.com/google_containers/alertmanager", "quay.io/prometheus/prometheus": "registry.cn-hangzhou.aliyuncs.com/google_containers/prometheus", - "k8s.gcr.io/ingress-nginx/kube-webhook-certgen": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-webhook-certgen", + "registry.k8s.io/ingress-nginx/kube-webhook-certgen": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-webhook-certgen", "gcr.io/k8s-minikube/minikube-ingress-dns": "registry.cn-hangzhou.aliyuncs.com/google_containers/minikube-ingress-dns", "gcr.io/google_containers/pause": "registry.cn-hangzhou.aliyuncs.com/google_containers/pause", - "k8s.gcr.io/metrics-server/metrics-server": "registry.cn-hangzhou.aliyuncs.com/google_containers/metrics-server", + "registry.k8s.io/metrics-server/metrics-server": "registry.cn-hangzhou.aliyuncs.com/google_containers/metrics-server", "gcr.io/google_containers/kube-registry-proxy": "registry.cn-hangzhou.aliyuncs.com/google_containers/kube-registry-proxy" } \ No newline at end of file diff --git a/deploy/addons/gpu/nvidia-driver-installer.yaml.tmpl b/deploy/addons/gpu/nvidia-driver-installer.yaml.tmpl index bc6b39c51b2d..83dc631750c7 100644 --- a/deploy/addons/gpu/nvidia-driver-installer.yaml.tmpl +++ b/deploy/addons/gpu/nvidia-driver-installer.yaml.tmpl @@ -77,5 +77,5 @@ spec: - name: root-mount mountPath: /root containers: - - image: "{{default "k8s.gcr.io" .ImageRepository}}/{{.Images.Pause}}" + - image: "{{default "registry.k8s.io" .ImageRepository}}/{{.Images.Pause}}" name: pause diff --git a/deploy/addons/gvisor/gvisor-config.toml b/deploy/addons/gvisor/gvisor-config.toml index bcf12b25b4f3..ba26b2d3eb52 100644 --- a/deploy/addons/gvisor/gvisor-config.toml +++ b/deploy/addons/gvisor/gvisor-config.toml @@ -29,7 +29,7 @@ oom_score = 0 stream_server_address = "" stream_server_port = "10010" enable_selinux = false - sandbox_image = "{{default "k8s.gcr.io" .ImageRepository}}/pause:3.1" + sandbox_image = "{{default "registry.k8s.io" .ImageRepository}}/pause:3.1" stats_collect_period = 10 systemd_cgroup = false enable_tls_streaming = false diff --git a/deploy/iso/minikube-iso/arch/aarch64/package/containerd-bin-aarch64/config.toml b/deploy/iso/minikube-iso/arch/aarch64/package/containerd-bin-aarch64/config.toml index 12b0d42eb7ac..873cee549d02 100644 --- a/deploy/iso/minikube-iso/arch/aarch64/package/containerd-bin-aarch64/config.toml +++ b/deploy/iso/minikube-iso/arch/aarch64/package/containerd-bin-aarch64/config.toml @@ -31,7 +31,7 @@ oom_score = 0 stream_server_address = "" stream_server_port = "10010" enable_selinux = false - sandbox_image = "k8s.gcr.io/pause:3.6" + sandbox_image = "registry.k8s.io/pause:3.6" stats_collect_period = 10 enable_tls_streaming = false max_container_log_line_size = 16384 diff --git a/deploy/iso/minikube-iso/arch/x86_64/package/containerd-bin/config.toml b/deploy/iso/minikube-iso/arch/x86_64/package/containerd-bin/config.toml index 12b0d42eb7ac..873cee549d02 100644 --- a/deploy/iso/minikube-iso/arch/x86_64/package/containerd-bin/config.toml +++ b/deploy/iso/minikube-iso/arch/x86_64/package/containerd-bin/config.toml @@ -31,7 +31,7 @@ oom_score = 0 stream_server_address = "" stream_server_port = "10010" enable_selinux = false - sandbox_image = "k8s.gcr.io/pause:3.6" + sandbox_image = "registry.k8s.io/pause:3.6" stats_collect_period = 10 enable_tls_streaming = false max_container_log_line_size = 16384 diff --git a/deploy/iso/minikube-iso/package/crio-bin/crio.conf b/deploy/iso/minikube-iso/package/crio-bin/crio.conf index 56133c41bc20..110d38473f3e 100644 --- a/deploy/iso/minikube-iso/package/crio-bin/crio.conf +++ b/deploy/iso/minikube-iso/package/crio-bin/crio.conf @@ -339,7 +339,7 @@ global_auth_file = "" # The image used to instantiate infra containers. # This option supports live configuration reload. -pause_image = "k8s.gcr.io/pause:3.2" +pause_image = "registry.k8s.io/pause:3.2" # The path to a file containing credentials specific for pulling the pause_image from # above. The file is similar to that of /var/lib/kubelet/config.json diff --git a/deploy/iso/minikube-iso/package/crio-bin/crio.conf.default b/deploy/iso/minikube-iso/package/crio-bin/crio.conf.default index 1a8ddfc3a568..1c557dcdc06b 100644 --- a/deploy/iso/minikube-iso/package/crio-bin/crio.conf.default +++ b/deploy/iso/minikube-iso/package/crio-bin/crio.conf.default @@ -339,7 +339,7 @@ global_auth_file = "" # The image used to instantiate infra containers. # This option supports live configuration reload. -pause_image = "k8s.gcr.io/pause:3.2" +pause_image = "registry.k8s.io/pause:3.2" # The path to a file containing credentials specific for pulling the pause_image from # above. The file is similar to that of /var/lib/kubelet/config.json diff --git a/deploy/kicbase/containerd.toml b/deploy/kicbase/containerd.toml index 12b0d42eb7ac..873cee549d02 100644 --- a/deploy/kicbase/containerd.toml +++ b/deploy/kicbase/containerd.toml @@ -31,7 +31,7 @@ oom_score = 0 stream_server_address = "" stream_server_port = "10010" enable_selinux = false - sandbox_image = "k8s.gcr.io/pause:3.6" + sandbox_image = "registry.k8s.io/pause:3.6" stats_collect_period = 10 enable_tls_streaming = false max_container_log_line_size = 16384 diff --git a/pkg/addons/addons.go b/pkg/addons/addons.go index aacdf43d5fee..ce7679a546f9 100644 --- a/pkg/addons/addons.go +++ b/pkg/addons/addons.go @@ -381,7 +381,7 @@ func supportLegacyIngress(addon *assets.Addon, cc config.ClusterConfig) error { "KubeWebhookCertgenPatch": "docker.io/jettech/kube-webhook-certgen:v1.5.1@sha256:950833e19ade18cd389d647efb88992a7cc077abedef343fa59e012d376d79b7", } addon.Registries = map[string]string{ - "IngressController": "k8s.gcr.io", + "IngressController": "registry.k8s.io", } return nil } diff --git a/pkg/minikube/assets/addons.go b/pkg/minikube/assets/addons.go index b2aab884de8a..012345f63f68 100644 --- a/pkg/minikube/assets/addons.go +++ b/pkg/minikube/assets/addons.go @@ -236,8 +236,8 @@ var Addons = map[string]*Addon{ "Alpine": "alpine:3.6@sha256:66790a2b79e1ea3e1dabac43990c54aca5d1ddf268d9a5a0285e4167c8b24475", "Kibana": "kibana/kibana:5.6.2@sha256:cd948a9bda4622f1437afc4a3e78be6c8c25fc62f40aa0376f3d690f2436568f", }, map[string]string{ - "Elasticsearch": "k8s.gcr.io", - "FluentdElasticsearch": "k8s.gcr.io", + "Elasticsearch": "registry.k8s.io", + "FluentdElasticsearch": "registry.k8s.io", "Kibana": "docker.elastic.co", "Alpine": "docker.io", }), @@ -436,8 +436,8 @@ var Addons = map[string]*Addon{ "NvidiaDriverInstaller": "minikube-nvidia-driver-installer:e2d9b43228decf5d6f7dce3f0a85d390f138fa01", "Pause": "pause:2.0@sha256:9ce5316f9752b8347484ab0f6778573af15524124d52b93230b9a0dcc987e73e", }, map[string]string{ - "NvidiaDriverInstaller": "k8s.gcr.io", - "Pause": "k8s.gcr.io", + "NvidiaDriverInstaller": "registry.k8s.io", + "Pause": "registry.k8s.io", }), "nvidia-gpu-device-plugin": NewAddon([]*BinAsset{ MustBinAsset(addons.NvidiaGpuDevicePluginAssets, @@ -448,7 +448,7 @@ var Addons = map[string]*Addon{ }, false, "nvidia-gpu-device-plugin", "3rd party (Nvidia)", "", "https://minikube.sigs.k8s.io/docs/tutorials/nvidia_gpu/", map[string]string{ "NvidiaDevicePlugin": "nvidia-gpu-device-plugin@sha256:4b036e8844920336fa48f36edeb7d4398f426d6a934ba022848deed2edbf09aa", }, map[string]string{ - "NvidiaDevicePlugin": "k8s.gcr.io", + "NvidiaDevicePlugin": "registry.k8s.io", }), "logviewer": NewAddon([]*BinAsset{ MustBinAsset(addons.LogviewerAssets, @@ -581,7 +581,7 @@ var Addons = map[string]*Addon{ "GCPAuthWebhook": "k8s-minikube/gcp-auth-webhook:v0.0.13@sha256:08a49cb7a588d81723b7e02c16082c75418b6e0a54cf2e44668bd77f79a41a40", }, map[string]string{ "GCPAuthWebhook": "gcr.io", - "KubeWebhookCertgen": "k8s.gcr.io", + "KubeWebhookCertgen": "registry.k8s.io", }), "volumesnapshots": NewAddon([]*BinAsset{ // make sure the order of apply. `csi-hostpath-snapshotclass` must be the first position, because it depends on `snapshot.storage.k8s.io_volumesnapshotclasses` @@ -619,7 +619,7 @@ var Addons = map[string]*Addon{ }, false, "volumesnapshots", "Kubernetes", "", "https://minikube.sigs.k8s.io/docs/tutorials/volume_snapshots_and_csi/", map[string]string{ "SnapshotController": "sig-storage/snapshot-controller:v4.0.0@sha256:00fcc441ea9f72899c25eed61d602272a2a58c5f0014332bdcb5ac24acef08e4", }, map[string]string{ - "SnapshotController": "k8s.gcr.io", + "SnapshotController": "registry.k8s.io", }), "csi-hostpath-driver": NewAddon([]*BinAsset{ MustBinAsset(addons.CsiHostpathDriverAssets, @@ -698,15 +698,15 @@ var Addons = map[string]*Addon{ "Snapshotter": "sig-storage/csi-snapshotter:v4.0.0@sha256:51f2dfde5bccac7854b3704689506aeecfb793328427b91115ba253a93e60782", "Provisioner": "sig-storage/csi-provisioner:v2.1.0@sha256:20c828075d1e36f679d6a91e905b0927141eef5e15be0c9a1ca4a6a0ed9313d2", }, map[string]string{ - "Attacher": "k8s.gcr.io", - "HostMonitorAgent": "k8s.gcr.io", - "HostMonitorController": "k8s.gcr.io", - "NodeDriverRegistrar": "k8s.gcr.io", - "HostPathPlugin": "k8s.gcr.io", - "LivenessProbe": "k8s.gcr.io", - "Resizer": "k8s.gcr.io", - "Snapshotter": "k8s.gcr.io", - "Provisioner": "k8s.gcr.io", + "Attacher": "registry.k8s.io", + "HostMonitorAgent": "registry.k8s.io", + "HostMonitorController": "registry.k8s.io", + "NodeDriverRegistrar": "registry.k8s.io", + "HostPathPlugin": "registry.k8s.io", + "LivenessProbe": "registry.k8s.io", + "Resizer": "registry.k8s.io", + "Snapshotter": "registry.k8s.io", + "Provisioner": "registry.k8s.io", }), "portainer": NewAddon([]*BinAsset{ MustBinAsset(addons.PortainerAssets, @@ -951,8 +951,8 @@ func GenerateTemplateData(addon *Addon, cc *config.ClusterConfig, netInfo Networ // tl;dr If the user specified a custom image remove the default registry // Without the line below, if you try to overwrite an image the default registry is still used in the templating // Example - image name: MetricsScraper, default registry: docker.io, default image: kubernetesui/metrics-scraper - // Passed on addon enable: --images=MetricsScraper=k8s.gcr.io/echoserver:1.4 - // Without this line the resulting image would be docker.io/k8s.gcr.io/echoserver:1.4 + // Passed on addon enable: --images=MetricsScraper=registry.k8s.io/echoserver:1.4 + // Without this line the resulting image would be docker.io/registry.k8s.io/echoserver:1.4 if _, ok := cc.CustomAddonImages[name]; ok { opts.Registries[name] = "" } diff --git a/pkg/minikube/bootstrapper/images/images.go b/pkg/minikube/bootstrapper/images/images.go index 72299e80dfb3..f4c8901b9fa4 100644 --- a/pkg/minikube/bootstrapper/images/images.go +++ b/pkg/minikube/bootstrapper/images/images.go @@ -34,7 +34,7 @@ import ( ) const ( - // builds a docker v2 repository API call in the format https://k8s.gcr.io/v2/coredns/coredns/tags/list + // builds a docker v2 repository API call in the format https://registry.k8s.io/v2/coredns/coredns/tags/list tagURLTemplate = "https://%s/v2/%s/tags/list" ) @@ -70,7 +70,7 @@ func componentImage(name string, v semver.Version, mirror string) string { return fmt.Sprintf("%s:v%s", path.Join(kubernetesRepo(mirror, v), name), v) } -// fixes 13136 by getting the latest image version from the k8s.gcr.io repository instead of hardcoded +// fixes 13136 by getting the latest image version from the registry.k8s.io repository instead of hardcoded func findLatestTagFromRepository(url string, lastKnownGood string) string { client := &http.Client{} errorMsg := fmt.Sprintf("Failed to get latest image version for %s, reverting to version %s.", url, lastKnownGood) diff --git a/pkg/minikube/bootstrapper/images/images_test.go b/pkg/minikube/bootstrapper/images/images_test.go index 67078d294a2d..96d97fa4d9fc 100644 --- a/pkg/minikube/bootstrapper/images/images_test.go +++ b/pkg/minikube/bootstrapper/images/images_test.go @@ -33,49 +33,49 @@ func TestEssentials(t *testing.T) { images []string }{ {"v1.18.0", strings.Split(strings.Trim(` -k8s.gcr.io/kube-apiserver:v1.18.0 -k8s.gcr.io/kube-controller-manager:v1.18.0 -k8s.gcr.io/kube-scheduler:v1.18.0 -k8s.gcr.io/kube-proxy:v1.18.0 -k8s.gcr.io/pause:3.2 -k8s.gcr.io/etcd:3.4.3-0 -k8s.gcr.io/coredns:1.6.7 +registry.k8s.io/kube-apiserver:v1.18.0 +registry.k8s.io/kube-controller-manager:v1.18.0 +registry.k8s.io/kube-scheduler:v1.18.0 +registry.k8s.io/kube-proxy:v1.18.0 +registry.k8s.io/pause:3.2 +registry.k8s.io/etcd:3.4.3-0 +registry.k8s.io/coredns:1.6.7 `, "\n"), "\n")}, {"v1.19.0", strings.Split(strings.Trim(` -k8s.gcr.io/kube-apiserver:v1.19.0 -k8s.gcr.io/kube-controller-manager:v1.19.0 -k8s.gcr.io/kube-scheduler:v1.19.0 -k8s.gcr.io/kube-proxy:v1.19.0 -k8s.gcr.io/pause:3.2 -k8s.gcr.io/etcd:3.4.9-1 -k8s.gcr.io/coredns:1.7.0 +registry.k8s.io/kube-apiserver:v1.19.0 +registry.k8s.io/kube-controller-manager:v1.19.0 +registry.k8s.io/kube-scheduler:v1.19.0 +registry.k8s.io/kube-proxy:v1.19.0 +registry.k8s.io/pause:3.2 +registry.k8s.io/etcd:3.4.9-1 +registry.k8s.io/coredns:1.7.0 `, "\n"), "\n")}, {"v1.20.0", strings.Split(strings.Trim(` -k8s.gcr.io/kube-apiserver:v1.20.0 -k8s.gcr.io/kube-controller-manager:v1.20.0 -k8s.gcr.io/kube-scheduler:v1.20.0 -k8s.gcr.io/kube-proxy:v1.20.0 -k8s.gcr.io/pause:3.2 -k8s.gcr.io/etcd:3.4.13-0 -k8s.gcr.io/coredns:1.7.0 +registry.k8s.io/kube-apiserver:v1.20.0 +registry.k8s.io/kube-controller-manager:v1.20.0 +registry.k8s.io/kube-scheduler:v1.20.0 +registry.k8s.io/kube-proxy:v1.20.0 +registry.k8s.io/pause:3.2 +registry.k8s.io/etcd:3.4.13-0 +registry.k8s.io/coredns:1.7.0 `, "\n"), "\n")}, {"v1.21.0", strings.Split(strings.Trim(` -k8s.gcr.io/kube-apiserver:v1.21.0 -k8s.gcr.io/kube-controller-manager:v1.21.0 -k8s.gcr.io/kube-scheduler:v1.21.0 -k8s.gcr.io/kube-proxy:v1.21.0 -k8s.gcr.io/pause:3.4.1 -k8s.gcr.io/etcd:3.4.13-0 -k8s.gcr.io/coredns/coredns:v1.8.0 +registry.k8s.io/kube-apiserver:v1.21.0 +registry.k8s.io/kube-controller-manager:v1.21.0 +registry.k8s.io/kube-scheduler:v1.21.0 +registry.k8s.io/kube-proxy:v1.21.0 +registry.k8s.io/pause:3.4.1 +registry.k8s.io/etcd:3.4.13-0 +registry.k8s.io/coredns/coredns:v1.8.0 `, "\n"), "\n")}, {"v1.22.0", strings.Split(strings.Trim(` -k8s.gcr.io/kube-apiserver:v1.22.0 -k8s.gcr.io/kube-controller-manager:v1.22.0 -k8s.gcr.io/kube-scheduler:v1.22.0 -k8s.gcr.io/kube-proxy:v1.22.0 -k8s.gcr.io/pause:3.5 -k8s.gcr.io/etcd:3.5.0-0 -k8s.gcr.io/coredns/coredns:v1.8.4 +registry.k8s.io/kube-apiserver:v1.22.0 +registry.k8s.io/kube-controller-manager:v1.22.0 +registry.k8s.io/kube-scheduler:v1.22.0 +registry.k8s.io/kube-proxy:v1.22.0 +registry.k8s.io/pause:3.5 +registry.k8s.io/etcd:3.5.0-0 +registry.k8s.io/coredns/coredns:v1.8.4 `, "\n"), "\n")}, } for _, tc := range testCases { @@ -85,7 +85,7 @@ k8s.gcr.io/coredns/coredns:v1.8.4 t.Fatal(err) } want := tc.images - got := essentials("k8s.gcr.io", v) + got := essentials("registry.k8s.io", v) if diff := cmp.Diff(want, got); diff != "" { t.Errorf("images mismatch (-want +got):\n%s", diff) } diff --git a/pkg/minikube/bootstrapper/images/kubeadm_test.go b/pkg/minikube/bootstrapper/images/kubeadm_test.go index cac1bc267279..7fb07394e3d7 100644 --- a/pkg/minikube/bootstrapper/images/kubeadm_test.go +++ b/pkg/minikube/bootstrapper/images/kubeadm_test.go @@ -75,43 +75,43 @@ func TestKubeadmImages(t *testing.T) { "mirror.k8s.io/k8s-minikube/storage-provisioner:" + version.GetStorageProvisionerVersion(), }}, {"v1.24.0", "", false, []string{ - "k8s.gcr.io/kube-proxy:v1.24.0", - "k8s.gcr.io/kube-scheduler:v1.24.0", - "k8s.gcr.io/kube-controller-manager:v1.24.0", - "k8s.gcr.io/kube-apiserver:v1.24.0", - "k8s.gcr.io/coredns/coredns:v1.8.6", - "k8s.gcr.io/etcd:3.5.3-0", - "k8s.gcr.io/pause:3.7", + "registry.k8s.io/kube-proxy:v1.24.0", + "registry.k8s.io/kube-scheduler:v1.24.0", + "registry.k8s.io/kube-controller-manager:v1.24.0", + "registry.k8s.io/kube-apiserver:v1.24.0", + "registry.k8s.io/coredns/coredns:v1.8.6", + "registry.k8s.io/etcd:3.5.3-0", + "registry.k8s.io/pause:3.7", "gcr.io/k8s-minikube/storage-provisioner:" + version.GetStorageProvisionerVersion(), }}, {"v1.23.0", "", false, []string{ - "k8s.gcr.io/kube-proxy:v1.23.0", - "k8s.gcr.io/kube-scheduler:v1.23.0", - "k8s.gcr.io/kube-controller-manager:v1.23.0", - "k8s.gcr.io/kube-apiserver:v1.23.0", - "k8s.gcr.io/coredns/coredns:v1.8.6", - "k8s.gcr.io/etcd:3.5.1-0", - "k8s.gcr.io/pause:3.6", + "registry.k8s.io/kube-proxy:v1.23.0", + "registry.k8s.io/kube-scheduler:v1.23.0", + "registry.k8s.io/kube-controller-manager:v1.23.0", + "registry.k8s.io/kube-apiserver:v1.23.0", + "registry.k8s.io/coredns/coredns:v1.8.6", + "registry.k8s.io/etcd:3.5.1-0", + "registry.k8s.io/pause:3.6", "gcr.io/k8s-minikube/storage-provisioner:" + version.GetStorageProvisionerVersion(), }}, {"v1.22.0", "", false, []string{ - "k8s.gcr.io/kube-proxy:v1.22.0", - "k8s.gcr.io/kube-scheduler:v1.22.0", - "k8s.gcr.io/kube-controller-manager:v1.22.0", - "k8s.gcr.io/kube-apiserver:v1.22.0", - "k8s.gcr.io/coredns/coredns:v1.8.4", - "k8s.gcr.io/etcd:3.5.0-0", - "k8s.gcr.io/pause:3.5", + "registry.k8s.io/kube-proxy:v1.22.0", + "registry.k8s.io/kube-scheduler:v1.22.0", + "registry.k8s.io/kube-controller-manager:v1.22.0", + "registry.k8s.io/kube-apiserver:v1.22.0", + "registry.k8s.io/coredns/coredns:v1.8.4", + "registry.k8s.io/etcd:3.5.0-0", + "registry.k8s.io/pause:3.5", "gcr.io/k8s-minikube/storage-provisioner:" + version.GetStorageProvisionerVersion(), }}, {"v1.16.0", "", false, []string{ - "k8s.gcr.io/kube-proxy:v1.16.0", - "k8s.gcr.io/kube-scheduler:v1.16.0", - "k8s.gcr.io/kube-controller-manager:v1.16.0", - "k8s.gcr.io/kube-apiserver:v1.16.0", - "k8s.gcr.io/coredns:1.6.2", - "k8s.gcr.io/etcd:3.3.15-0", - "k8s.gcr.io/pause:3.1", + "registry.k8s.io/kube-proxy:v1.16.0", + "registry.k8s.io/kube-scheduler:v1.16.0", + "registry.k8s.io/kube-controller-manager:v1.16.0", + "registry.k8s.io/kube-apiserver:v1.16.0", + "registry.k8s.io/coredns:1.6.2", + "registry.k8s.io/etcd:3.3.15-0", + "registry.k8s.io/pause:3.1", "gcr.io/k8s-minikube/storage-provisioner:" + version.GetStorageProvisionerVersion(), }}, {"v1.11.0", "", true, nil}, diff --git a/pkg/minikube/bootstrapper/images/repo.go b/pkg/minikube/bootstrapper/images/repo.go index 956d9d13f8f1..d3b2e8566cc5 100644 --- a/pkg/minikube/bootstrapper/images/repo.go +++ b/pkg/minikube/bootstrapper/images/repo.go @@ -21,7 +21,7 @@ import ( ) // OldDefaultKubernetesRepo is the old default Kubernetes repository -const OldDefaultKubernetesRepo = "k8s.gcr.io" +const OldDefaultKubernetesRepo = "registry.k8s.io" // NewDefaultKubernetesRepo is the new default Kubernetes repository const NewDefaultKubernetesRepo = "registry.k8s.io" diff --git a/pkg/minikube/config/types.go b/pkg/minikube/config/types.go index 7aec572b8abf..32481b3b4a65 100644 --- a/pkg/minikube/config/types.go +++ b/pkg/minikube/config/types.go @@ -76,7 +76,7 @@ type ClusterConfig struct { KubernetesConfig KubernetesConfig Nodes []Node Addons map[string]bool - CustomAddonImages map[string]string // Maps image names to the image to use for addons. e.g. Dashboard -> k8s.gcr.io/echoserver:1.4 makes dashboard addon use echoserver for its Dashboard deployment. + CustomAddonImages map[string]string // Maps image names to the image to use for addons. e.g. Dashboard -> registry.k8s.io/echoserver:1.4 makes dashboard addon use echoserver for its Dashboard deployment. CustomAddonRegistries map[string]string // Maps image names to the registry to use for addons. See CustomAddonImages for example. VerifyComponents map[string]bool // map of components to verify and wait for after start. StartHostTimeout time.Duration diff --git a/pkg/minikube/image/cache.go b/pkg/minikube/image/cache.go index aa1b9e8be8ad..e25d0f09bcb6 100644 --- a/pkg/minikube/image/cache.go +++ b/pkg/minikube/image/cache.go @@ -61,8 +61,8 @@ func DeleteFromCacheDir(images []string) error { // SaveToDir will cache images on the host // // The cache directory currently caches images using the imagename_tag -// For example, k8s.gcr.io/kube-addon-manager:v6.5 would be -// stored at $CACHE_DIR/k8s.gcr.io/kube-addon-manager_v6.5 +// For example, registry.k8s.io/kube-addon-manager:v6.5 would be +// stored at $CACHE_DIR/registry.k8s.io/kube-addon-manager_v6.5 func SaveToDir(images []string, cacheDir string, overwrite bool) error { var g errgroup.Group for _, image := range images { diff --git a/pkg/minikube/reason/match_test.go b/pkg/minikube/reason/match_test.go index 72c19dd984e3..d049a8c43434 100644 --- a/pkg/minikube/reason/match_test.go +++ b/pkg/minikube/reason/match_test.go @@ -53,7 +53,7 @@ VBoxManage.exe: error: Details: code E_FAIL (0x80004005), component MachineWrap, {4913, "linux", "PR_KVM_CREATE_BUSY", `Unable to start VM: create: Error creating machine: Error in driver during machine creation: error creating VM: virError(Code=1, Domain=10, Message='internal error: process exited while connecting to monitor: ioctl(KVM_CREATE_VM) failed: 16 Device or resource busy`}, {5950, "linux", "PR_KVM_ISO_PERMISSION", `Retriable failure: create: Error creating machine: Error in driver during machine creation: error creating VM: virError(Code=1, Domain=10, Message='internal error: qemu unexpectedly closed the monitor: 2019-11-19T16:08:16.757609Z qemu-kvm: -drive file=/home/lnicotra/.minikube/machines/minikube/boot2docker.iso,format=raw,if=none,id=drive-scsi0-0-0-2,readonly=on: could not open disk image /home/lnicotra/.minikube/machines/minikube/boot2docker.iso: Could not open '/home/lnicotra/.minikube/machines/minikube/boot2docker.iso': Permission denied'`}, {5836, "", "SVC_OPEN_NOT_FOUND", `Error opening service: Service kubernetes-bootcamp was not found in "default" namespace. You may select another namespace by using 'minikube service kubernetes-bootcamp -n : Temporary Error: Error getting service kubernetes-bootcamp: services "kubernetes-bootcamp" not found`}, - {3898, "", "INET_PULL_TIMEOUT", `[ERROR ImagePull]: failed to pull image k8s.gcr.io/kube-controller-manager:v1.17.0: output: Error response from daemon: Get https://k8s.gcr.io/v2/: net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)`}, + {3898, "", "INET_PULL_TIMEOUT", `[ERROR ImagePull]: failed to pull image registry.k8s.io/kube-controller-manager:v1.17.0: output: Error response from daemon: Get https://registry.k8s.io/v2/: net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)`}, {6079, "darwin", "PR_HYPERKIT_CRASHED", `Error creating machine: Error in driver during machine creation: hyperkit crashed! command line:`}, {5636, "linux", "INET_DEFAULT_ROUTE", `Unable to get VM IP address: unable to select an IP from default routes.`}, {6087, "", "GUEST_DOES_NOT_EXIST", `Error getting host status: state: machine does not exist`}, diff --git a/site/content/en/docs/contrib/tests.en.md b/site/content/en/docs/contrib/tests.en.md index 4025bac8df00..3f2401b111e3 100644 --- a/site/content/en/docs/contrib/tests.en.md +++ b/site/content/en/docs/contrib/tests.en.md @@ -294,7 +294,7 @@ Steps: asserts basic "service" command functionality Steps: -- Create a new `k8s.gcr.io/echoserver` deployment +- Create a new `registry.k8s.io/echoserver` deployment - Run `minikube service list` to make sure the newly created service is correctly listed in the output - Run `minikube service` with `--https --url` to make sure the HTTPS endpoint URL of the service is printed - Run `minikube service` with `--url --format={{.IP}}` to make sure the IP address of the service is printed @@ -303,7 +303,7 @@ Steps: #### validateServiceCmdConnect Steps: -- Create a new `k8s.gcr.io/echoserver` deployment +- Create a new `registry.k8s.io/echoserver` deployment - Run `minikube service` with a regular `--url` to make sure the HTTP endpoint URL of the service is printed - Make sure we can hit the endpoint URL with an HTTP GET request diff --git a/site/content/en/docs/handbook/addons/custom-images.md b/site/content/en/docs/handbook/addons/custom-images.md index a66c44ee092b..d1549288a202 100644 --- a/site/content/en/docs/handbook/addons/custom-images.md +++ b/site/content/en/docs/handbook/addons/custom-images.md @@ -19,8 +19,8 @@ minikube addons images efk |----------------------|------------------------------|-------------------| | IMAGE NAME | DEFAULT IMAGE | DEFAULT REGISTRY | |----------------------|------------------------------|-------------------| -| Elasticsearch | elasticsearch:v5.6.2 | k8s.gcr.io | -| FluentdElasticsearch | fluentd-elasticsearch:v2.0.2 | k8s.gcr.io | +| Elasticsearch | elasticsearch:v5.6.2 | registry.k8s.io | +| FluentdElasticsearch | fluentd-elasticsearch:v2.0.2 | registry.k8s.io | | Alpine | alpine:3.6 | | | Kibana | kibana/kibana:5.6.2 | docker.elastic.co | |----------------------|------------------------------|-------------------| @@ -31,7 +31,7 @@ An empty registry means the image is stored locally or default registry `docker. The `IMAGE NAME` column is used to customize the corresponding image and registry. -Assume we have a private registry at `192.168.10.2:5555` to replace `k8s.gcr.io` and a locally built Kibana called `kibana/kibana:5.6.2-custom`. +Assume we have a private registry at `192.168.10.2:5555` to replace `registry.k8s.io` and a locally built Kibana called `kibana/kibana:5.6.2-custom`. We could load local images to minikube by: diff --git a/site/content/en/docs/handbook/offline.md b/site/content/en/docs/handbook/offline.md index d9238520fb13..b15670ee1bdc 100644 --- a/site/content/en/docs/handbook/offline.md +++ b/site/content/en/docs/handbook/offline.md @@ -27,18 +27,18 @@ For offline use on other hosts, one can copy the contents of `~/.minikube/cache` ```text cache/iso/minikube-v1.0.0.iso cache/images/gcr.io/k8s-minikube/storage-provisioner_v1.8.1 -cache/images/k8s.gcr.io/k8s-dns-sidecar-amd64_1.14.13 -cache/images/k8s.gcr.io/k8s-dns-dnsmasq-nanny-amd64_1.14.13 -cache/images/k8s.gcr.io/kubernetes-dashboard-amd64_v1.10.1 -cache/images/k8s.gcr.io/kube-scheduler_v1.14.0 -cache/images/k8s.gcr.io/coredns_1.3.1 -cache/images/k8s.gcr.io/kube-controller-manager_v1.14.0 -cache/images/k8s.gcr.io/kube-apiserver_v1.14.0 -cache/images/k8s.gcr.io/pause_3.1 -cache/images/k8s.gcr.io/etcd_3.3.10 -cache/images/k8s.gcr.io/kube-addon-manager_v9.0 -cache/images/k8s.gcr.io/k8s-dns-kube-dns-amd64_1.14.13 -cache/images/k8s.gcr.io/kube-proxy_v1.14.0 +cache/images/registry.k8s.io/k8s-dns-sidecar-amd64_1.14.13 +cache/images/registry.k8s.io/k8s-dns-dnsmasq-nanny-amd64_1.14.13 +cache/images/registry.k8s.io/kubernetes-dashboard-amd64_v1.10.1 +cache/images/registry.k8s.io/kube-scheduler_v1.14.0 +cache/images/registry.k8s.io/coredns_1.3.1 +cache/images/registry.k8s.io/kube-controller-manager_v1.14.0 +cache/images/registry.k8s.io/kube-apiserver_v1.14.0 +cache/images/registry.k8s.io/pause_3.1 +cache/images/registry.k8s.io/etcd_3.3.10 +cache/images/registry.k8s.io/kube-addon-manager_v9.0 +cache/images/registry.k8s.io/k8s-dns-kube-dns-amd64_1.14.13 +cache/images/registry.k8s.io/kube-proxy_v1.14.0 cache/v1.14.0/kubeadm cache/v1.14.0/kubelet ``` diff --git a/site/content/en/docs/handbook/pushing.md b/site/content/en/docs/handbook/pushing.md index 8e46c444dbb5..c55837455b65 100644 --- a/site/content/en/docs/handbook/pushing.md +++ b/site/content/en/docs/handbook/pushing.md @@ -392,7 +392,7 @@ buildctl --addr unix://buildkitd.sock build \ --frontend=dockerfile.v0 \ --local context=. \ --local dockerfile=. \ - --output type=image,name=k8s.gcr.io/username/imagename:latest + --output type=image,name=registry.k8s.io/username/imagename:latest ``` Now you can 'build' against the storage inside minikube. which is instantly accessible to kubernetes cluster. diff --git a/site/content/en/docs/handbook/vpn_and_proxy.md b/site/content/en/docs/handbook/vpn_and_proxy.md index 9042feefb6a0..9e19afa6367a 100644 --- a/site/content/en/docs/handbook/vpn_and_proxy.md +++ b/site/content/en/docs/handbook/vpn_and_proxy.md @@ -72,8 +72,8 @@ This error indicates that the host:port combination defined by HTTPS_PROXY or HT ```text Unable to pull images, which may be OK: -failed to pull image "k8s.gcr.io/kube-apiserver:v1.13.3": output: Error response from daemon: -Get https://k8s.gcr.io/v2/: net/http: request canceled while waiting for connection +failed to pull image "registry.k8s.io/kube-apiserver:v1.13.3": output: Error response from daemon: +Get https://registry.k8s.io/v2/: net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) ``` @@ -82,9 +82,9 @@ This error indicates that the container runtime running within the VM does not h #### x509: certificate signed by unknown authority ```text -[ERROR ImagePull]: failed to pull image k8s.gcr.io/kube-apiserver:v1.13.3: +[ERROR ImagePull]: failed to pull image registry.k8s.io/kube-apiserver:v1.13.3: output: Error response from daemon: -Get https://k8s.gcr.io/v2/: x509: certificate signed by unknown authority +Get https://registry.k8s.io/v2/: x509: certificate signed by unknown authority ``` This is because minikube VM is stuck behind a proxy that rewrites HTTPS responses to contain its own TLS certificate. The [solution](https://github.com/kubernetes/minikube/issues/3613#issuecomment-461034222) is to install the proxy certificate into a location that is copied to the VM at startup, so that it can be validated. diff --git a/test/integration/functional_test.go b/test/integration/functional_test.go index a221a73d49fa..3265662d52dc 100644 --- a/test/integration/functional_test.go +++ b/test/integration/functional_test.go @@ -276,7 +276,7 @@ func runImageList(ctx context.Context, t *testing.T, profile, testName, format, func expectedImageFormat(format string) []string { return []string{ - fmt.Sprintf(format, "k8s.gcr.io/pause"), + fmt.Sprintf(format, "registry.k8s.io/pause"), fmt.Sprintf(format, "registry.k8s.io/kube-apiserver"), } } @@ -1038,7 +1038,7 @@ func validateCacheCmd(ctx context.Context, t *testing.T, profile string) { // docs: Run `minikube cache add` and make sure we can add a remote image to the cache t.Run("add_remote", func(t *testing.T) { - for _, img := range []string{"k8s.gcr.io/pause:3.1", "k8s.gcr.io/pause:3.3", "k8s.gcr.io/pause:latest"} { + for _, img := range []string{"registry.k8s.io/pause:3.1", "registry.k8s.io/pause:3.3", "registry.k8s.io/pause:latest"} { rr, err := Run(t, exec.CommandContext(ctx, Target(), "-p", profile, "cache", "add", img)) if err != nil { t.Errorf("failed to 'cache add' remote image %q. args %q err %v", img, rr.Command(), err) @@ -1091,10 +1091,10 @@ func validateCacheCmd(ctx context.Context, t *testing.T, profile string) { }) // docs: Run `minikube cache delete` and make sure we can delete an image from the cache - t.Run("delete_k8s.gcr.io/pause:3.3", func(t *testing.T) { - rr, err := Run(t, exec.CommandContext(ctx, Target(), "cache", "delete", "k8s.gcr.io/pause:3.3")) + t.Run("delete_registry.k8s.io/pause:3.3", func(t *testing.T) { + rr, err := Run(t, exec.CommandContext(ctx, Target(), "cache", "delete", "registry.k8s.io/pause:3.3")) if err != nil { - t.Errorf("failed to delete image k8s.gcr.io/pause:3.3 from cache. args %q: %v", rr.Command(), err) + t.Errorf("failed to delete image registry.k8s.io/pause:3.3 from cache. args %q: %v", rr.Command(), err) } }) @@ -1104,11 +1104,11 @@ func validateCacheCmd(ctx context.Context, t *testing.T, profile string) { if err != nil { t.Errorf("failed to do cache list. args %q: %v", rr.Command(), err) } - if !strings.Contains(rr.Output(), "k8s.gcr.io/pause") { - t.Errorf("expected 'cache list' output to include 'k8s.gcr.io/pause' but got: ***%s***", rr.Output()) + if !strings.Contains(rr.Output(), "registry.k8s.io/pause") { + t.Errorf("expected 'cache list' output to include 'registry.k8s.io/pause' but got: ***%s***", rr.Output()) } - if strings.Contains(rr.Output(), "k8s.gcr.io/pause:3.3") { - t.Errorf("expected 'cache list' output not to include k8s.gcr.io/pause:3.3 but got: ***%s***", rr.Output()) + if strings.Contains(rr.Output(), "registry.k8s.io/pause:3.3") { + t.Errorf("expected 'cache list' output not to include registry.k8s.io/pause:3.3 but got: ***%s***", rr.Output()) } }) @@ -1126,7 +1126,7 @@ func validateCacheCmd(ctx context.Context, t *testing.T, profile string) { // docs: Delete an image from minikube node and run `minikube cache reload` to make sure the image is brought back correctly t.Run("cache_reload", func(t *testing.T) { // deleting image inside minikube node manually and expecting reload to bring it back - img := "k8s.gcr.io/pause:latest" + img := "registry.k8s.io/pause:latest" // deleting image inside minikube node manually var binary string @@ -1161,7 +1161,7 @@ func validateCacheCmd(ctx context.Context, t *testing.T, profile string) { // delete will clean up the cached images since they are global and all other tests will load it for no reason t.Run("delete", func(t *testing.T) { - for _, img := range []string{"k8s.gcr.io/pause:3.1", "k8s.gcr.io/pause:latest"} { + for _, img := range []string{"registry.k8s.io/pause:3.1", "registry.k8s.io/pause:latest"} { rr, err := Run(t, exec.CommandContext(ctx, Target(), "cache", "delete", img)) if err != nil { t.Errorf("failed to delete %s from cache. args %q: %v", img, rr.Command(), err) @@ -1425,12 +1425,12 @@ func validateServiceCmd(ctx context.Context, t *testing.T, profile string) { var rr *RunResult var err error - // docs: Create a new `k8s.gcr.io/echoserver` deployment - // k8s.gcr.io/echoserver is not multi-arch + // docs: Create a new `registry.k8s.io/echoserver` deployment + // registry.k8s.io/echoserver is not multi-arch if arm64Platform() { - rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "create", "deployment", "hello-node", "--image=k8s.gcr.io/echoserver-arm:1.8")) + rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "create", "deployment", "hello-node", "--image=registry.k8s.io/echoserver-arm:1.8")) } else { - rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "create", "deployment", "hello-node", "--image=k8s.gcr.io/echoserver:1.8")) + rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "create", "deployment", "hello-node", "--image=registry.k8s.io/echoserver:1.8")) } if err != nil { @@ -1551,12 +1551,12 @@ func validateServiceCmdConnect(ctx context.Context, t *testing.T, profile string var rr *RunResult var err error - // docs: Create a new `k8s.gcr.io/echoserver` deployment - // k8s.gcr.io/echoserver is not multi-arch + // docs: Create a new `registry.k8s.io/echoserver` deployment + // registry.k8s.io/echoserver is not multi-arch if arm64Platform() { - rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "create", "deployment", "hello-node-connect", "--image=k8s.gcr.io/echoserver-arm:1.8")) + rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "create", "deployment", "hello-node-connect", "--image=registry.k8s.io/echoserver-arm:1.8")) } else { - rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "create", "deployment", "hello-node-connect", "--image=k8s.gcr.io/echoserver:1.8")) + rr, err = Run(t, exec.CommandContext(ctx, "kubectl", "--context", profile, "create", "deployment", "hello-node-connect", "--image=registry.k8s.io/echoserver:1.8")) } if err != nil { diff --git a/test/integration/start_stop_delete_test.go b/test/integration/start_stop_delete_test.go index e8435c3e3d58..280e27206c0e 100644 --- a/test/integration/start_stop_delete_test.go +++ b/test/integration/start_stop_delete_test.go @@ -202,7 +202,7 @@ func validateEnableAddonWhileActive(ctx context.Context, t *testing.T, profile s defer PostMortemLogs(t, profile) // Enable an addon to assert it requests the correct image. - rr, err := Run(t, exec.CommandContext(ctx, Target(), "addons", "enable", "metrics-server", "-p", profile, "--images=MetricsServer=k8s.gcr.io/echoserver:1.4", "--registries=MetricsServer=fake.domain")) + rr, err := Run(t, exec.CommandContext(ctx, Target(), "addons", "enable", "metrics-server", "-p", profile, "--images=MetricsServer=registry.k8s.io/echoserver:1.4", "--registries=MetricsServer=fake.domain")) if err != nil { t.Errorf("failed to enable an addon post-stop. args %q: %v", rr.Command(), err) } @@ -217,8 +217,8 @@ func validateEnableAddonWhileActive(ctx context.Context, t *testing.T, profile s t.Errorf("failed to get info on auto-pause deployments. args %q: %v", rr.Command(), err) } deploymentInfo := rr.Stdout.String() - if !strings.Contains(deploymentInfo, " fake.domain/k8s.gcr.io/echoserver:1.4") { - t.Errorf("addon did not load correct image. Expected to contain \" fake.domain/k8s.gcr.io/echoserver:1.4\". Addon deployment info: %s", deploymentInfo) + if !strings.Contains(deploymentInfo, " fake.domain/registry.k8s.io/echoserver:1.4") { + t.Errorf("addon did not load correct image. Expected to contain \" fake.domain/registry.k8s.io/echoserver:1.4\". Addon deployment info: %s", deploymentInfo) } } @@ -243,7 +243,7 @@ func validateEnableAddonAfterStop(ctx context.Context, t *testing.T, profile str } // Enable an addon to assert it comes up afterwards - rr, err := Run(t, exec.CommandContext(ctx, Target(), "addons", "enable", "dashboard", "-p", profile, "--images=MetricsScraper=k8s.gcr.io/echoserver:1.4")) + rr, err := Run(t, exec.CommandContext(ctx, Target(), "addons", "enable", "dashboard", "-p", profile, "--images=MetricsScraper=registry.k8s.io/echoserver:1.4")) if err != nil { t.Errorf("failed to enable an addon post-stop. args %q: %v", rr.Command(), err) } @@ -293,8 +293,8 @@ func validateAddonAfterStop(ctx context.Context, t *testing.T, profile string, t t.Errorf("failed to get info on kubernetes-dashboard deployments. args %q: %v", rr.Command(), err) } deploymentInfo := rr.Stdout.String() - if !strings.Contains(deploymentInfo, " k8s.gcr.io/echoserver:1.4") { - t.Errorf("addon did not load correct image. Expected to contain \" k8s.gcr.io/echoserver:1.4\". Addon deployment info: %s", deploymentInfo) + if !strings.Contains(deploymentInfo, " registry.k8s.io/echoserver:1.4") { + t.Errorf("addon did not load correct image. Expected to contain \" registry.k8s.io/echoserver:1.4\". Addon deployment info: %s", deploymentInfo) } } @@ -464,7 +464,7 @@ func defaultImage(name string) bool { if strings.Contains(name, ":latest") { return false } - if strings.Contains(name, "k8s.gcr.io") || strings.Contains(name, "registry.k8s.io") || strings.Contains(name, "kubernetesui") || strings.Contains(name, "storage-provisioner") { + if strings.Contains(name, "registry.k8s.io") || strings.Contains(name, "registry.k8s.io") || strings.Contains(name, "kubernetesui") || strings.Contains(name, "storage-provisioner") { return true } return false diff --git a/test/integration/testdata/multinodes/multinode-pod-dns-test.yaml b/test/integration/testdata/multinodes/multinode-pod-dns-test.yaml index c778c28d04b2..28b3d3c927c5 100644 --- a/test/integration/testdata/multinodes/multinode-pod-dns-test.yaml +++ b/test/integration/testdata/multinodes/multinode-pod-dns-test.yaml @@ -18,7 +18,7 @@ spec: - name: busybox # flaky nslookup in busybox versions newer than 1.28: # https://github.com/docker-library/busybox/issues/48 - # note: k8s.gcr.io/e2e-test-images/agnhost:2.32 + # note: registry.k8s.io/e2e-test-images/agnhost:2.32 # has similar issues (ie, resolves but returns exit code 1) image: gcr.io/k8s-minikube/busybox:1.28 command: