=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:332: registry stabilized in 1.868167ms
addons_test.go:334: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-t5sv4" [2f41b6f7-f293-467f-8215-b24af50ec8ba] Running
addons_test.go:334: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 6.002866149s
addons_test.go:337: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-z9ss9" [29edb9a3-341b-486a-8045-5546e8911d8c] Running
addons_test.go:337: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.002559219s
addons_test.go:342: (dbg) Run: kubectl --context addons-118348 delete po -l run=registry-test --now
addons_test.go:347: (dbg) Run: kubectl --context addons-118348 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:347: (dbg) Non-zero exit: kubectl --context addons-118348 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.070828696s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:349: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-118348 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:353: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:361: (dbg) Run: out/minikube-linux-amd64 -p addons-118348 ip
2024/09/17 08:51:37 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:390: (dbg) Run: out/minikube-linux-amd64 -p addons-118348 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-118348
helpers_test.go:235: (dbg) docker inspect addons-118348:
-- stdout --
[
{
"Id": "446a733f1f3a5dd9a1b27181ad91f934e43da3a3c6e2c831d7491d51c849bb20",
"Created": "2024-09-17T08:38:35.768048295Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 16890,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-17T08:38:35.895667879Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:20d492278eed428d119466f58713403332b5d2ac1db7c6863f797e2406f2b671",
"ResolvConfPath": "/var/lib/docker/containers/446a733f1f3a5dd9a1b27181ad91f934e43da3a3c6e2c831d7491d51c849bb20/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/446a733f1f3a5dd9a1b27181ad91f934e43da3a3c6e2c831d7491d51c849bb20/hostname",
"HostsPath": "/var/lib/docker/containers/446a733f1f3a5dd9a1b27181ad91f934e43da3a3c6e2c831d7491d51c849bb20/hosts",
"LogPath": "/var/lib/docker/containers/446a733f1f3a5dd9a1b27181ad91f934e43da3a3c6e2c831d7491d51c849bb20/446a733f1f3a5dd9a1b27181ad91f934e43da3a3c6e2c831d7491d51c849bb20-json.log",
"Name": "/addons-118348",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-118348:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-118348",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/11013b4f49cb49fefe3bf53f86efce26167ad1c783a7d916f52cc097b212e611-init/diff:/var/lib/docker/overlay2/7da256a43b3639c4f92f439ecfea8165b0571eba2633ca08d3d0447ef408406e/diff",
"MergedDir": "/var/lib/docker/overlay2/11013b4f49cb49fefe3bf53f86efce26167ad1c783a7d916f52cc097b212e611/merged",
"UpperDir": "/var/lib/docker/overlay2/11013b4f49cb49fefe3bf53f86efce26167ad1c783a7d916f52cc097b212e611/diff",
"WorkDir": "/var/lib/docker/overlay2/11013b4f49cb49fefe3bf53f86efce26167ad1c783a7d916f52cc097b212e611/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-118348",
"Source": "/var/lib/docker/volumes/addons-118348/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-118348",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-118348",
"name.minikube.sigs.k8s.io": "addons-118348",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "b41e9b99330a4fce7d668f0abcfb9cadff52dc936b35b5a72804519a3a3abdf8",
"SandboxKey": "/var/run/docker/netns/b41e9b99330a",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-118348": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "309e07f873c6e539134ea98773c9a158e538a4844e94270780ae606901223d6d",
"EndpointID": "2455298c01d72a47ee361ac85b347908cd7a74c9e987885310b37ff2d8893fab",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-118348",
"446a733f1f3a"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-118348 -n addons-118348
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-118348 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-docker-450334 | download-docker-450334 | jenkins | v1.34.0 | 17 Sep 24 08:38 UTC | 17 Sep 24 08:38 UTC |
| start | --download-only -p | binary-mirror-437165 | jenkins | v1.34.0 | 17 Sep 24 08:38 UTC | |
| | binary-mirror-437165 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:39083 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-437165 | binary-mirror-437165 | jenkins | v1.34.0 | 17 Sep 24 08:38 UTC | 17 Sep 24 08:38 UTC |
| addons | disable dashboard -p | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:38 UTC | |
| | addons-118348 | | | | | |
| addons | enable dashboard -p | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:38 UTC | |
| | addons-118348 | | | | | |
| start | -p addons-118348 --wait=true | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:38 UTC | 17 Sep 24 08:41 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | addons-118348 addons disable | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:42 UTC | 17 Sep 24 08:42 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:50 UTC | 17 Sep 24 08:50 UTC |
| | addons-118348 | | | | | |
| addons | addons-118348 addons | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:50 UTC | 17 Sep 24 08:50 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-118348 addons disable | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:50 UTC | 17 Sep 24 08:50 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ssh | addons-118348 ssh cat | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:50 UTC | 17 Sep 24 08:50 UTC |
| | /opt/local-path-provisioner/pvc-55d397ea-86e9-4f5a-ae73-814393eaf4d2_default_test-pvc/file1 | | | | | |
| addons | addons-118348 addons disable | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:50 UTC | 17 Sep 24 08:51 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-118348 addons disable | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:50 UTC | 17 Sep 24 08:50 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:50 UTC | 17 Sep 24 08:50 UTC |
| | -p addons-118348 | | | | | |
| addons | addons-118348 addons | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-118348 ssh curl -s | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-118348 ip | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| addons | addons-118348 addons | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-118348 addons disable | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-118348 addons disable | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| addons | disable cloud-spanner -p | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| | addons-118348 | | | | | |
| addons | enable headlamp | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| | -p addons-118348 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-118348 addons disable | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ip | addons-118348 ip | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| addons | addons-118348 addons disable | addons-118348 | jenkins | v1.34.0 | 17 Sep 24 08:51 UTC | 17 Sep 24 08:51 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/17 08:38:12
Running on machine: ubuntu-20-agent-14
Binary: Built with gc go1.23.0 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0917 08:38:12.306438 16153 out.go:345] Setting OutFile to fd 1 ...
I0917 08:38:12.306560 16153 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0917 08:38:12.306575 16153 out.go:358] Setting ErrFile to fd 2...
I0917 08:38:12.306580 16153 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0917 08:38:12.306768 16153 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19648-8091/.minikube/bin
I0917 08:38:12.307427 16153 out.go:352] Setting JSON to false
I0917 08:38:12.308259 16153 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":1243,"bootTime":1726561049,"procs":173,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1069-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0917 08:38:12.308349 16153 start.go:139] virtualization: kvm guest
I0917 08:38:12.310565 16153 out.go:177] * [addons-118348] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0917 08:38:12.312108 16153 out.go:177] - MINIKUBE_LOCATION=19648
I0917 08:38:12.312112 16153 notify.go:220] Checking for updates...
I0917 08:38:12.314632 16153 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0917 08:38:12.315863 16153 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19648-8091/kubeconfig
I0917 08:38:12.317139 16153 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19648-8091/.minikube
I0917 08:38:12.318165 16153 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0917 08:38:12.319245 16153 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0917 08:38:12.320401 16153 driver.go:394] Setting default libvirt URI to qemu:///system
I0917 08:38:12.342410 16153 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
I0917 08:38:12.342525 16153 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0917 08:38:12.390925 16153 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-17 08:38:12.381794766 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0917 08:38:12.391047 16153 docker.go:318] overlay module found
I0917 08:38:12.392959 16153 out.go:177] * Using the docker driver based on user configuration
I0917 08:38:12.394095 16153 start.go:297] selected driver: docker
I0917 08:38:12.394106 16153 start.go:901] validating driver "docker" against <nil>
I0917 08:38:12.394117 16153 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0917 08:38:12.394855 16153 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0917 08:38:12.441823 16153 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-17 08:38:12.432994306 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0917 08:38:12.441969 16153 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0917 08:38:12.442210 16153 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0917 08:38:12.443921 16153 out.go:177] * Using Docker driver with root privileges
I0917 08:38:12.445517 16153 cni.go:84] Creating CNI manager for ""
I0917 08:38:12.445586 16153 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0917 08:38:12.445599 16153 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0917 08:38:12.445675 16153 start.go:340] cluster config:
{Name:addons-118348 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-118348 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0917 08:38:12.447372 16153 out.go:177] * Starting "addons-118348" primary control-plane node in "addons-118348" cluster
I0917 08:38:12.448900 16153 cache.go:121] Beginning downloading kic base image for docker with docker
I0917 08:38:12.450200 16153 out.go:177] * Pulling base image v0.0.45-1726358845-19644 ...
I0917 08:38:12.451247 16153 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0917 08:38:12.451287 16153 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 in local docker daemon
I0917 08:38:12.451293 16153 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19648-8091/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4
I0917 08:38:12.451311 16153 cache.go:56] Caching tarball of preloaded images
I0917 08:38:12.451404 16153 preload.go:172] Found /home/jenkins/minikube-integration/19648-8091/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0917 08:38:12.451416 16153 cache.go:59] Finished verifying existence of preloaded tar for v1.31.1 on docker
I0917 08:38:12.451733 16153 profile.go:143] Saving config to /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/config.json ...
I0917 08:38:12.451757 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/config.json: {Name:mk154b312e5e7a9628d069a344fd855bd4470df3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:12.468389 16153 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 to local cache
I0917 08:38:12.468512 16153 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 in local cache directory
I0917 08:38:12.468528 16153 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 in local cache directory, skipping pull
I0917 08:38:12.468533 16153 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 exists in cache, skipping pull
I0917 08:38:12.468543 16153 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 as a tarball
I0917 08:38:12.468550 16153 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 from local cache
I0917 08:38:24.257565 16153 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 from cached tarball
I0917 08:38:24.257614 16153 cache.go:194] Successfully downloaded all kic artifacts
I0917 08:38:24.257655 16153 start.go:360] acquireMachinesLock for addons-118348: {Name:mk092a2f95e180c254fe3c2f3a6c594e1014ed24 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0917 08:38:24.257751 16153 start.go:364] duration metric: took 78.173µs to acquireMachinesLock for "addons-118348"
I0917 08:38:24.257773 16153 start.go:93] Provisioning new machine with config: &{Name:addons-118348 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-118348 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0917 08:38:24.257850 16153 start.go:125] createHost starting for "" (driver="docker")
I0917 08:38:24.259806 16153 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0917 08:38:24.260072 16153 start.go:159] libmachine.API.Create for "addons-118348" (driver="docker")
I0917 08:38:24.260103 16153 client.go:168] LocalClient.Create starting
I0917 08:38:24.260229 16153 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19648-8091/.minikube/certs/ca.pem
I0917 08:38:24.371715 16153 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19648-8091/.minikube/certs/cert.pem
I0917 08:38:24.469429 16153 cli_runner.go:164] Run: docker network inspect addons-118348 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0917 08:38:24.485011 16153 cli_runner.go:211] docker network inspect addons-118348 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0917 08:38:24.485077 16153 network_create.go:284] running [docker network inspect addons-118348] to gather additional debugging logs...
I0917 08:38:24.485097 16153 cli_runner.go:164] Run: docker network inspect addons-118348
W0917 08:38:24.501204 16153 cli_runner.go:211] docker network inspect addons-118348 returned with exit code 1
I0917 08:38:24.501259 16153 network_create.go:287] error running [docker network inspect addons-118348]: docker network inspect addons-118348: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-118348 not found
I0917 08:38:24.501279 16153 network_create.go:289] output of [docker network inspect addons-118348]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-118348 not found
** /stderr **
I0917 08:38:24.501395 16153 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0917 08:38:24.517794 16153 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001a9e8e0}
I0917 08:38:24.517836 16153 network_create.go:124] attempt to create docker network addons-118348 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0917 08:38:24.517882 16153 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-118348 addons-118348
I0917 08:38:24.579581 16153 network_create.go:108] docker network addons-118348 192.168.49.0/24 created
I0917 08:38:24.579615 16153 kic.go:121] calculated static IP "192.168.49.2" for the "addons-118348" container
I0917 08:38:24.579678 16153 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0917 08:38:24.595318 16153 cli_runner.go:164] Run: docker volume create addons-118348 --label name.minikube.sigs.k8s.io=addons-118348 --label created_by.minikube.sigs.k8s.io=true
I0917 08:38:24.613336 16153 oci.go:103] Successfully created a docker volume addons-118348
I0917 08:38:24.613403 16153 cli_runner.go:164] Run: docker run --rm --name addons-118348-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-118348 --entrypoint /usr/bin/test -v addons-118348:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -d /var/lib
I0917 08:38:31.794593 16153 cli_runner.go:217] Completed: docker run --rm --name addons-118348-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-118348 --entrypoint /usr/bin/test -v addons-118348:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -d /var/lib: (7.181153302s)
I0917 08:38:31.794624 16153 oci.go:107] Successfully prepared a docker volume addons-118348
I0917 08:38:31.794649 16153 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0917 08:38:31.794672 16153 kic.go:194] Starting extracting preloaded images to volume ...
I0917 08:38:31.794743 16153 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19648-8091/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-118348:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -I lz4 -xf /preloaded.tar -C /extractDir
I0917 08:38:35.705599 16153 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19648-8091/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-118348:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -I lz4 -xf /preloaded.tar -C /extractDir: (3.910803356s)
I0917 08:38:35.705630 16153 kic.go:203] duration metric: took 3.91095417s to extract preloaded images to volume ...
W0917 08:38:35.705757 16153 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0917 08:38:35.705848 16153 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0917 08:38:35.753901 16153 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-118348 --name addons-118348 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-118348 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-118348 --network addons-118348 --ip 192.168.49.2 --volume addons-118348:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0
I0917 08:38:36.078035 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Running}}
I0917 08:38:36.095751 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:36.114738 16153 cli_runner.go:164] Run: docker exec addons-118348 stat /var/lib/dpkg/alternatives/iptables
I0917 08:38:36.157132 16153 oci.go:144] the created container "addons-118348" has a running status.
I0917 08:38:36.157159 16153 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa...
I0917 08:38:36.263732 16153 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0917 08:38:36.283327 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:36.300181 16153 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0917 08:38:36.300205 16153 kic_runner.go:114] Args: [docker exec --privileged addons-118348 chown docker:docker /home/docker/.ssh/authorized_keys]
I0917 08:38:36.345923 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:36.368316 16153 machine.go:93] provisionDockerMachine start ...
I0917 08:38:36.368412 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:36.387467 16153 main.go:141] libmachine: Using SSH client type: native
I0917 08:38:36.387725 16153 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 08:38:36.387739 16153 main.go:141] libmachine: About to run SSH command:
hostname
I0917 08:38:36.388456 16153 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: read tcp 127.0.0.1:48316->127.0.0.1:32768: read: connection reset by peer
I0917 08:38:39.515816 16153 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-118348
I0917 08:38:39.515843 16153 ubuntu.go:169] provisioning hostname "addons-118348"
I0917 08:38:39.515895 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:39.533039 16153 main.go:141] libmachine: Using SSH client type: native
I0917 08:38:39.533222 16153 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 08:38:39.533234 16153 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-118348 && echo "addons-118348" | sudo tee /etc/hostname
I0917 08:38:39.674459 16153 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-118348
I0917 08:38:39.674536 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:39.691347 16153 main.go:141] libmachine: Using SSH client type: native
I0917 08:38:39.691516 16153 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 08:38:39.691533 16153 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-118348' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-118348/g' /etc/hosts;
else
echo '127.0.1.1 addons-118348' | sudo tee -a /etc/hosts;
fi
fi
I0917 08:38:39.820463 16153 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0917 08:38:39.820489 16153 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19648-8091/.minikube CaCertPath:/home/jenkins/minikube-integration/19648-8091/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19648-8091/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19648-8091/.minikube}
I0917 08:38:39.820518 16153 ubuntu.go:177] setting up certificates
I0917 08:38:39.820527 16153 provision.go:84] configureAuth start
I0917 08:38:39.820576 16153 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-118348
I0917 08:38:39.837257 16153 provision.go:143] copyHostCerts
I0917 08:38:39.837344 16153 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19648-8091/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19648-8091/.minikube/ca.pem (1082 bytes)
I0917 08:38:39.837475 16153 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19648-8091/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19648-8091/.minikube/cert.pem (1123 bytes)
I0917 08:38:39.837542 16153 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19648-8091/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19648-8091/.minikube/key.pem (1679 bytes)
I0917 08:38:39.837597 16153 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19648-8091/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19648-8091/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19648-8091/.minikube/certs/ca-key.pem org=jenkins.addons-118348 san=[127.0.0.1 192.168.49.2 addons-118348 localhost minikube]
I0917 08:38:39.937274 16153 provision.go:177] copyRemoteCerts
I0917 08:38:39.937328 16153 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0917 08:38:39.937383 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:39.954152 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:40.048990 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I0917 08:38:40.070119 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0917 08:38:40.091179 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0917 08:38:40.112372 16153 provision.go:87] duration metric: took 291.834318ms to configureAuth
I0917 08:38:40.112399 16153 ubuntu.go:193] setting minikube options for container-runtime
I0917 08:38:40.112586 16153 config.go:182] Loaded profile config "addons-118348": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0917 08:38:40.112636 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:40.129243 16153 main.go:141] libmachine: Using SSH client type: native
I0917 08:38:40.129433 16153 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 08:38:40.129449 16153 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0917 08:38:40.260919 16153 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0917 08:38:40.260943 16153 ubuntu.go:71] root file system type: overlay
I0917 08:38:40.261042 16153 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0917 08:38:40.261108 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:40.278136 16153 main.go:141] libmachine: Using SSH client type: native
I0917 08:38:40.278326 16153 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 08:38:40.278414 16153 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0917 08:38:40.422782 16153 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0917 08:38:40.422874 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:40.439499 16153 main.go:141] libmachine: Using SSH client type: native
I0917 08:38:40.439699 16153 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 08:38:40.439725 16153 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0917 08:38:41.117254 16153 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-09-06 12:06:41.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-17 08:38:40.418272262 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0917 08:38:41.117287 16153 machine.go:96] duration metric: took 4.748944653s to provisionDockerMachine
I0917 08:38:41.117299 16153 client.go:171] duration metric: took 16.857187768s to LocalClient.Create
I0917 08:38:41.117314 16153 start.go:167] duration metric: took 16.857244012s to libmachine.API.Create "addons-118348"
I0917 08:38:41.117322 16153 start.go:293] postStartSetup for "addons-118348" (driver="docker")
I0917 08:38:41.117332 16153 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0917 08:38:41.117389 16153 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0917 08:38:41.117427 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:41.133609 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:41.228941 16153 ssh_runner.go:195] Run: cat /etc/os-release
I0917 08:38:41.231827 16153 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0917 08:38:41.231854 16153 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0917 08:38:41.231862 16153 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0917 08:38:41.231868 16153 info.go:137] Remote host: Ubuntu 22.04.4 LTS
I0917 08:38:41.231878 16153 filesync.go:126] Scanning /home/jenkins/minikube-integration/19648-8091/.minikube/addons for local assets ...
I0917 08:38:41.231936 16153 filesync.go:126] Scanning /home/jenkins/minikube-integration/19648-8091/.minikube/files for local assets ...
I0917 08:38:41.231958 16153 start.go:296] duration metric: took 114.630368ms for postStartSetup
I0917 08:38:41.232229 16153 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-118348
I0917 08:38:41.248320 16153 profile.go:143] Saving config to /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/config.json ...
I0917 08:38:41.248642 16153 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0917 08:38:41.248737 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:41.264623 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:41.353099 16153 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0917 08:38:41.356911 16153 start.go:128] duration metric: took 17.099047972s to createHost
I0917 08:38:41.356931 16153 start.go:83] releasing machines lock for "addons-118348", held for 17.099169362s
I0917 08:38:41.356981 16153 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-118348
I0917 08:38:41.373005 16153 ssh_runner.go:195] Run: cat /version.json
I0917 08:38:41.373050 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:41.373101 16153 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0917 08:38:41.373174 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:41.390001 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:41.390261 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:41.480120 16153 ssh_runner.go:195] Run: systemctl --version
I0917 08:38:41.572448 16153 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0917 08:38:41.576513 16153 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0917 08:38:41.598233 16153 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0917 08:38:41.598292 16153 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0917 08:38:41.623399 16153 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0917 08:38:41.623421 16153 start.go:495] detecting cgroup driver to use...
I0917 08:38:41.623451 16153 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0917 08:38:41.623552 16153 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0917 08:38:41.637945 16153 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0917 08:38:41.646767 16153 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0917 08:38:41.655342 16153 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0917 08:38:41.655406 16153 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0917 08:38:41.663993 16153 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0917 08:38:41.672564 16153 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0917 08:38:41.681274 16153 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0917 08:38:41.689756 16153 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0917 08:38:41.697773 16153 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0917 08:38:41.706361 16153 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0917 08:38:41.715393 16153 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0917 08:38:41.724141 16153 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0917 08:38:41.731534 16153 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0917 08:38:41.738995 16153 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 08:38:41.812796 16153 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0917 08:38:41.899219 16153 start.go:495] detecting cgroup driver to use...
I0917 08:38:41.899268 16153 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0917 08:38:41.899327 16153 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0917 08:38:41.910462 16153 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0917 08:38:41.910520 16153 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0917 08:38:41.920463 16153 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0917 08:38:41.934444 16153 ssh_runner.go:195] Run: which cri-dockerd
I0917 08:38:41.937542 16153 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0917 08:38:41.945753 16153 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0917 08:38:41.961435 16153 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0917 08:38:42.044624 16153 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0917 08:38:42.141996 16153 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0917 08:38:42.142133 16153 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0917 08:38:42.158856 16153 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 08:38:42.233124 16153 ssh_runner.go:195] Run: sudo systemctl restart docker
I0917 08:38:42.479936 16153 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0917 08:38:42.491044 16153 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0917 08:38:42.501365 16153 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0917 08:38:42.577057 16153 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0917 08:38:42.651787 16153 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 08:38:42.725849 16153 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0917 08:38:42.737777 16153 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0917 08:38:42.747432 16153 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 08:38:42.817904 16153 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0917 08:38:42.877001 16153 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0917 08:38:42.877082 16153 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0917 08:38:42.880348 16153 start.go:563] Will wait 60s for crictl version
I0917 08:38:42.880397 16153 ssh_runner.go:195] Run: which crictl
I0917 08:38:42.883417 16153 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0917 08:38:42.913967 16153 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.2.1
RuntimeApiVersion: v1
I0917 08:38:42.914058 16153 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0917 08:38:42.935849 16153 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0917 08:38:42.960600 16153 out.go:235] * Preparing Kubernetes v1.31.1 on Docker 27.2.1 ...
I0917 08:38:42.960694 16153 cli_runner.go:164] Run: docker network inspect addons-118348 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0917 08:38:42.976623 16153 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0917 08:38:42.979950 16153 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0917 08:38:42.989610 16153 kubeadm.go:883] updating cluster {Name:addons-118348 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-118348 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0917 08:38:42.989714 16153 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0917 08:38:42.989759 16153 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0917 08:38:43.007240 16153 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0917 08:38:43.007263 16153 docker.go:615] Images already preloaded, skipping extraction
I0917 08:38:43.007317 16153 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0917 08:38:43.025100 16153 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0917 08:38:43.025130 16153 cache_images.go:84] Images are preloaded, skipping loading
I0917 08:38:43.025140 16153 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.1 docker true true} ...
I0917 08:38:43.025253 16153 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-118348 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.1 ClusterName:addons-118348 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0917 08:38:43.025321 16153 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0917 08:38:43.068745 16153 cni.go:84] Creating CNI manager for ""
I0917 08:38:43.068775 16153 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0917 08:38:43.068787 16153 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0917 08:38:43.068812 16153 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-118348 NodeName:addons-118348 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0917 08:38:43.069004 16153 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-118348"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0917 08:38:43.069072 16153 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.1
I0917 08:38:43.076969 16153 binaries.go:44] Found k8s binaries, skipping transfer
I0917 08:38:43.077034 16153 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0917 08:38:43.084770 16153 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0917 08:38:43.100226 16153 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0917 08:38:43.115895 16153 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0917 08:38:43.131409 16153 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0917 08:38:43.134406 16153 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0917 08:38:43.144055 16153 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 08:38:43.218261 16153 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0917 08:38:43.230466 16153 certs.go:68] Setting up /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348 for IP: 192.168.49.2
I0917 08:38:43.230486 16153 certs.go:194] generating shared ca certs ...
I0917 08:38:43.230501 16153 certs.go:226] acquiring lock for ca certs: {Name:mk3225b0343d01afc54a59e630093b9fbe48964d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.230607 16153 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19648-8091/.minikube/ca.key
I0917 08:38:43.322611 16153 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19648-8091/.minikube/ca.crt ...
I0917 08:38:43.322639 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/ca.crt: {Name:mkcfc5e560f028dd268362f8000159a9120a365d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.322799 16153 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19648-8091/.minikube/ca.key ...
I0917 08:38:43.322808 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/ca.key: {Name:mk32d8e669bcada4813d2823d783c47616f0c295 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.322877 16153 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19648-8091/.minikube/proxy-client-ca.key
I0917 08:38:43.507926 16153 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19648-8091/.minikube/proxy-client-ca.crt ...
I0917 08:38:43.507954 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/proxy-client-ca.crt: {Name:mk5f73d943ca8acd15c7a26e7442fcfedfc6ebde Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.508117 16153 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19648-8091/.minikube/proxy-client-ca.key ...
I0917 08:38:43.508127 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/proxy-client-ca.key: {Name:mk720d7c6f3866942f6fe9b02c1755e0caf82391 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.508194 16153 certs.go:256] generating profile certs ...
I0917 08:38:43.508250 16153 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/client.key
I0917 08:38:43.508273 16153 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/client.crt with IP's: []
I0917 08:38:43.642041 16153 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/client.crt ...
I0917 08:38:43.642074 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/client.crt: {Name:mk2c49dedb09e040b5b5ecc4dae40d1925164c9b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.642274 16153 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/client.key ...
I0917 08:38:43.642287 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/client.key: {Name:mke78e69e6ae7bfab20a29e99161db5897f8c80a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.642381 16153 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.key.29a37712
I0917 08:38:43.642403 16153 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.crt.29a37712 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0917 08:38:43.801001 16153 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.crt.29a37712 ...
I0917 08:38:43.801032 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.crt.29a37712: {Name:mk277794572cd2b73424aaa5162dbcb3fb7932df Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.801216 16153 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.key.29a37712 ...
I0917 08:38:43.801231 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.key.29a37712: {Name:mkbc8142b378fbc7e450ffe2722399e15683315d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.801342 16153 certs.go:381] copying /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.crt.29a37712 -> /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.crt
I0917 08:38:43.801414 16153 certs.go:385] copying /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.key.29a37712 -> /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.key
I0917 08:38:43.801459 16153 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/proxy-client.key
I0917 08:38:43.801473 16153 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/proxy-client.crt with IP's: []
I0917 08:38:43.886609 16153 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/proxy-client.crt ...
I0917 08:38:43.886641 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/proxy-client.crt: {Name:mkfaada3eea6c9655d10a1543f1611779c4f9ba4 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.886822 16153 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/proxy-client.key ...
I0917 08:38:43.886835 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/proxy-client.key: {Name:mkfcbb701aaeaf6dda621c8f4d2e47391158cd34 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:43.887026 16153 certs.go:484] found cert: /home/jenkins/minikube-integration/19648-8091/.minikube/certs/ca-key.pem (1679 bytes)
I0917 08:38:43.887060 16153 certs.go:484] found cert: /home/jenkins/minikube-integration/19648-8091/.minikube/certs/ca.pem (1082 bytes)
I0917 08:38:43.887082 16153 certs.go:484] found cert: /home/jenkins/minikube-integration/19648-8091/.minikube/certs/cert.pem (1123 bytes)
I0917 08:38:43.887103 16153 certs.go:484] found cert: /home/jenkins/minikube-integration/19648-8091/.minikube/certs/key.pem (1679 bytes)
I0917 08:38:43.887655 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0917 08:38:43.909591 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0917 08:38:43.931201 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0917 08:38:43.952558 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0917 08:38:43.973925 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0917 08:38:43.995669 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0917 08:38:44.017073 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0917 08:38:44.039192 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/profiles/addons-118348/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0917 08:38:44.060394 16153 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-8091/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0917 08:38:44.082554 16153 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0917 08:38:44.098902 16153 ssh_runner.go:195] Run: openssl version
I0917 08:38:44.103793 16153 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0917 08:38:44.113170 16153 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0917 08:38:44.116248 16153 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 17 08:38 /usr/share/ca-certificates/minikubeCA.pem
I0917 08:38:44.116308 16153 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0917 08:38:44.122465 16153 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0917 08:38:44.130892 16153 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0917 08:38:44.133922 16153 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0917 08:38:44.133978 16153 kubeadm.go:392] StartCluster: {Name:addons-118348 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-118348 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0917 08:38:44.134072 16153 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0917 08:38:44.150773 16153 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0917 08:38:44.158746 16153 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0917 08:38:44.166506 16153 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0917 08:38:44.166565 16153 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0917 08:38:44.174171 16153 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0917 08:38:44.174187 16153 kubeadm.go:157] found existing configuration files:
I0917 08:38:44.174221 16153 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0917 08:38:44.181699 16153 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0917 08:38:44.181750 16153 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0917 08:38:44.189264 16153 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0917 08:38:44.197011 16153 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0917 08:38:44.197064 16153 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0917 08:38:44.205191 16153 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0917 08:38:44.212779 16153 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0917 08:38:44.212828 16153 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0917 08:38:44.220338 16153 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0917 08:38:44.228286 16153 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0917 08:38:44.228339 16153 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0917 08:38:44.236089 16153 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0917 08:38:44.270065 16153 kubeadm.go:310] [init] Using Kubernetes version: v1.31.1
I0917 08:38:44.270145 16153 kubeadm.go:310] [preflight] Running pre-flight checks
I0917 08:38:44.289454 16153 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0917 08:38:44.289543 16153 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1069-gcp[0m
I0917 08:38:44.289588 16153 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0917 08:38:44.289646 16153 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0917 08:38:44.289708 16153 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0917 08:38:44.289768 16153 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0917 08:38:44.289830 16153 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0917 08:38:44.289893 16153 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0917 08:38:44.289956 16153 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0917 08:38:44.290003 16153 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0917 08:38:44.290069 16153 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0917 08:38:44.290131 16153 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0917 08:38:44.338058 16153 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0917 08:38:44.338174 16153 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0917 08:38:44.338306 16153 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0917 08:38:44.348322 16153 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0917 08:38:44.352031 16153 out.go:235] - Generating certificates and keys ...
I0917 08:38:44.352132 16153 kubeadm.go:310] [certs] Using existing ca certificate authority
I0917 08:38:44.352203 16153 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0917 08:38:44.392301 16153 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0917 08:38:44.500596 16153 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0917 08:38:44.656895 16153 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0917 08:38:44.858169 16153 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0917 08:38:45.200292 16153 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0917 08:38:45.200427 16153 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-118348 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0917 08:38:45.292627 16153 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0917 08:38:45.292828 16153 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-118348 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0917 08:38:45.930538 16153 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0917 08:38:46.146360 16153 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0917 08:38:46.259979 16153 kubeadm.go:310] [certs] Generating "sa" key and public key
I0917 08:38:46.260095 16153 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0917 08:38:46.502352 16153 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0917 08:38:46.552326 16153 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0917 08:38:46.679109 16153 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0917 08:38:47.179140 16153 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0917 08:38:47.339579 16153 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0917 08:38:47.340068 16153 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0917 08:38:47.342651 16153 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0917 08:38:47.344585 16153 out.go:235] - Booting up control plane ...
I0917 08:38:47.344679 16153 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0917 08:38:47.344778 16153 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0917 08:38:47.344855 16153 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0917 08:38:47.353131 16153 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0917 08:38:47.357910 16153 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0917 08:38:47.357974 16153 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0917 08:38:47.440992 16153 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0917 08:38:47.441123 16153 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0917 08:38:48.442298 16153 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.001379947s
I0917 08:38:48.442387 16153 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0917 08:38:52.944454 16153 kubeadm.go:310] [api-check] The API server is healthy after 4.502184668s
I0917 08:38:52.956023 16153 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0917 08:38:52.967288 16153 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0917 08:38:52.984977 16153 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0917 08:38:52.985204 16153 kubeadm.go:310] [mark-control-plane] Marking the node addons-118348 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0917 08:38:52.993123 16153 kubeadm.go:310] [bootstrap-token] Using token: 6dqjjf.f0nxv5t7pz6fpmqz
I0917 08:38:52.994681 16153 out.go:235] - Configuring RBAC rules ...
I0917 08:38:52.994786 16153 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0917 08:38:52.998509 16153 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0917 08:38:53.007331 16153 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0917 08:38:53.010197 16153 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0917 08:38:53.012942 16153 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0917 08:38:53.015405 16153 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0917 08:38:53.350764 16153 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0917 08:38:53.796981 16153 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0917 08:38:54.352177 16153 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0917 08:38:54.353007 16153 kubeadm.go:310]
I0917 08:38:54.353080 16153 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0917 08:38:54.353112 16153 kubeadm.go:310]
I0917 08:38:54.353241 16153 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0917 08:38:54.353252 16153 kubeadm.go:310]
I0917 08:38:54.353310 16153 kubeadm.go:310] mkdir -p $HOME/.kube
I0917 08:38:54.353422 16153 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0917 08:38:54.353499 16153 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0917 08:38:54.353514 16153 kubeadm.go:310]
I0917 08:38:54.353589 16153 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0917 08:38:54.353599 16153 kubeadm.go:310]
I0917 08:38:54.353661 16153 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0917 08:38:54.353670 16153 kubeadm.go:310]
I0917 08:38:54.353739 16153 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0917 08:38:54.353850 16153 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0917 08:38:54.353948 16153 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0917 08:38:54.353957 16153 kubeadm.go:310]
I0917 08:38:54.354079 16153 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0917 08:38:54.354160 16153 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0917 08:38:54.354166 16153 kubeadm.go:310]
I0917 08:38:54.354258 16153 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 6dqjjf.f0nxv5t7pz6fpmqz \
I0917 08:38:54.354377 16153 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:f39726bcd9567e166cfea3356daada82e147516a8cdf8435266f67c8416305f5 \
I0917 08:38:54.354403 16153 kubeadm.go:310] --control-plane
I0917 08:38:54.354411 16153 kubeadm.go:310]
I0917 08:38:54.354516 16153 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0917 08:38:54.354525 16153 kubeadm.go:310]
I0917 08:38:54.354639 16153 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 6dqjjf.f0nxv5t7pz6fpmqz \
I0917 08:38:54.354771 16153 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:f39726bcd9567e166cfea3356daada82e147516a8cdf8435266f67c8416305f5
I0917 08:38:54.356650 16153 kubeadm.go:310] W0917 08:38:44.267501 1920 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0917 08:38:54.356957 16153 kubeadm.go:310] W0917 08:38:44.268119 1920 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0917 08:38:54.357203 16153 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1069-gcp\n", err: exit status 1
I0917 08:38:54.357399 16153 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0917 08:38:54.357419 16153 cni.go:84] Creating CNI manager for ""
I0917 08:38:54.357431 16153 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0917 08:38:54.359202 16153 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0917 08:38:54.360435 16153 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0917 08:38:54.368904 16153 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0917 08:38:54.385480 16153 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0917 08:38:54.385617 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:54.385646 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-118348 minikube.k8s.io/updated_at=2024_09_17T08_38_54_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=9256ba43b41ea130fa48757ddb8d93db00574f61 minikube.k8s.io/name=addons-118348 minikube.k8s.io/primary=true
I0917 08:38:54.476282 16153 ops.go:34] apiserver oom_adj: -16
I0917 08:38:54.476389 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:54.976985 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:55.476465 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:55.976820 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:56.477361 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:56.976761 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:57.477019 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:57.977010 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:58.476795 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:58.976866 16153 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 08:38:59.039660 16153 kubeadm.go:1113] duration metric: took 4.654098927s to wait for elevateKubeSystemPrivileges
I0917 08:38:59.039698 16153 kubeadm.go:394] duration metric: took 14.905723458s to StartCluster
I0917 08:38:59.039720 16153 settings.go:142] acquiring lock: {Name:mk862a5c46e81240a806a4c66f0c2efde4cdc586 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:59.039845 16153 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19648-8091/kubeconfig
I0917 08:38:59.040228 16153 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-8091/kubeconfig: {Name:mk91a6d671e6a7ab453b1c24cade89fd7db9b782 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 08:38:59.040431 16153 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0917 08:38:59.040449 16153 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0917 08:38:59.040517 16153 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0917 08:38:59.040628 16153 addons.go:69] Setting yakd=true in profile "addons-118348"
I0917 08:38:59.040639 16153 addons.go:69] Setting cloud-spanner=true in profile "addons-118348"
I0917 08:38:59.040633 16153 addons.go:69] Setting default-storageclass=true in profile "addons-118348"
I0917 08:38:59.040653 16153 addons.go:234] Setting addon cloud-spanner=true in "addons-118348"
I0917 08:38:59.040661 16153 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-118348"
I0917 08:38:59.040662 16153 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-118348"
I0917 08:38:59.040666 16153 addons.go:69] Setting metrics-server=true in profile "addons-118348"
I0917 08:38:59.040678 16153 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-118348"
I0917 08:38:59.040706 16153 addons.go:234] Setting addon metrics-server=true in "addons-118348"
I0917 08:38:59.040708 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.040701 16153 addons.go:69] Setting ingress=true in profile "addons-118348"
I0917 08:38:59.040719 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.040728 16153 addons.go:234] Setting addon ingress=true in "addons-118348"
I0917 08:38:59.040736 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.040742 16153 config.go:182] Loaded profile config "addons-118348": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0917 08:38:59.040771 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.040791 16153 addons.go:69] Setting gcp-auth=true in profile "addons-118348"
I0917 08:38:59.040807 16153 mustload.go:65] Loading cluster: addons-118348
I0917 08:38:59.040835 16153 addons.go:69] Setting registry=true in profile "addons-118348"
I0917 08:38:59.040845 16153 addons.go:234] Setting addon registry=true in "addons-118348"
I0917 08:38:59.040862 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.040942 16153 config.go:182] Loaded profile config "addons-118348": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0917 08:38:59.041053 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.041172 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.041182 16153 addons.go:69] Setting helm-tiller=true in profile "addons-118348"
I0917 08:38:59.041192 16153 addons.go:234] Setting addon helm-tiller=true in "addons-118348"
I0917 08:38:59.041210 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.041234 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.041266 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.041286 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.041405 16153 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-118348"
I0917 08:38:59.041421 16153 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-118348"
I0917 08:38:59.041494 16153 addons.go:69] Setting volcano=true in profile "addons-118348"
I0917 08:38:59.041514 16153 addons.go:234] Setting addon volcano=true in "addons-118348"
I0917 08:38:59.041540 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.041622 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.041714 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.040654 16153 addons.go:234] Setting addon yakd=true in "addons-118348"
I0917 08:38:59.042186 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.042204 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.042499 16153 addons.go:69] Setting volumesnapshots=true in profile "addons-118348"
I0917 08:38:59.042516 16153 addons.go:234] Setting addon volumesnapshots=true in "addons-118348"
I0917 08:38:59.042537 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.043112 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.041173 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.043893 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.045790 16153 out.go:177] * Verifying Kubernetes components...
I0917 08:38:59.043913 16153 addons.go:69] Setting storage-provisioner=true in profile "addons-118348"
I0917 08:38:59.047858 16153 addons.go:234] Setting addon storage-provisioner=true in "addons-118348"
I0917 08:38:59.047909 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.043953 16153 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-118348"
I0917 08:38:59.048133 16153 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-118348"
I0917 08:38:59.043991 16153 addons.go:69] Setting ingress-dns=true in profile "addons-118348"
I0917 08:38:59.048159 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.048180 16153 addons.go:234] Setting addon ingress-dns=true in "addons-118348"
I0917 08:38:59.048272 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.044043 16153 addons.go:69] Setting inspektor-gadget=true in profile "addons-118348"
I0917 08:38:59.048365 16153 addons.go:234] Setting addon inspektor-gadget=true in "addons-118348"
I0917 08:38:59.048443 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.048600 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.048822 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.050189 16153 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 08:38:59.074360 16153 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0917 08:38:59.075732 16153 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0917 08:38:59.075793 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0917 08:38:59.075884 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.077062 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.077063 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.077536 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.080536 16153 addons.go:234] Setting addon default-storageclass=true in "addons-118348"
I0917 08:38:59.080584 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.081293 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.083350 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.086783 16153 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0917 08:38:59.088333 16153 out.go:177] - Using image docker.io/registry:2.8.3
I0917 08:38:59.088665 16153 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0917 08:38:59.088705 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0917 08:38:59.088759 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.098760 16153 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0917 08:38:59.103323 16153 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0917 08:38:59.103354 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0917 08:38:59.103413 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.116027 16153 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0917 08:38:59.117768 16153 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0917 08:38:59.119003 16153 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0917 08:38:59.120374 16153 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0917 08:38:59.120395 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0917 08:38:59.120456 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.128859 16153 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0917 08:38:59.128896 16153 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0917 08:38:59.128964 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.131084 16153 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0917 08:38:59.132421 16153 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0917 08:38:59.132449 16153 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0917 08:38:59.132518 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.137018 16153 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0917 08:38:59.138509 16153 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0917 08:38:59.138537 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0917 08:38:59.138604 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.141184 16153 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0917 08:38:59.142583 16153 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0917 08:38:59.142586 16153 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0917 08:38:59.142697 16153 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0917 08:38:59.142775 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.143969 16153 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0917 08:38:59.143989 16153 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0917 08:38:59.144042 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.148643 16153 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0917 08:38:59.150999 16153 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0917 08:38:59.152301 16153 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0917 08:38:59.155205 16153 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0917 08:38:59.155228 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0917 08:38:59.155280 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.158119 16153 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-118348"
I0917 08:38:59.158168 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:38:59.158644 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:38:59.160443 16153 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0917 08:38:59.161682 16153 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0917 08:38:59.161810 16153 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0917 08:38:59.161835 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0917 08:38:59.161907 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.163771 16153 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0917 08:38:59.164887 16153 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0917 08:38:59.165013 16153 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0917 08:38:59.165233 16153 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0917 08:38:59.165314 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.167546 16153 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0917 08:38:59.171149 16153 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0917 08:38:59.189535 16153 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0917 08:38:59.190143 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.195295 16153 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0917 08:38:59.195514 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.196473 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.196572 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.196864 16153 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0917 08:38:59.198382 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.198512 16153 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0917 08:38:59.198532 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0917 08:38:59.198598 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.198762 16153 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0917 08:38:59.200219 16153 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0917 08:38:59.201421 16153 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0917 08:38:59.201441 16153 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0917 08:38:59.201512 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.219412 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.222531 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.228414 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.228414 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.235142 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.236852 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.239129 16153 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0917 08:38:59.239975 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.241610 16153 out.go:177] - Using image docker.io/busybox:stable
I0917 08:38:59.242843 16153 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0917 08:38:59.242858 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0917 08:38:59.242905 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:38:59.247419 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.248960 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:38:59.260062 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
W0917 08:38:59.276011 16153 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0917 08:38:59.276054 16153 retry.go:31] will retry after 166.260019ms: ssh: handshake failed: EOF
I0917 08:38:59.505713 16153 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0917 08:38:59.505732 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0917 08:38:59.573417 16153 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0917 08:38:59.573441 16153 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0917 08:38:59.600205 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0917 08:38:59.674847 16153 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0917 08:38:59.674933 16153 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0917 08:38:59.676026 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0917 08:38:59.693102 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0917 08:38:59.695785 16153 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0917 08:38:59.695858 16153 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0917 08:38:59.781745 16153 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0917 08:38:59.781779 16153 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0917 08:38:59.789613 16153 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0917 08:38:59.789704 16153 ssh_runner.go:362] scp helm-tiller/helm-tiller-rbac.yaml --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0917 08:38:59.791786 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0917 08:38:59.873323 16153 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0917 08:38:59.873413 16153 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0917 08:38:59.882197 16153 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0917 08:38:59.882284 16153 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0917 08:38:59.890265 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0917 08:38:59.974407 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0917 08:38:59.976672 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0917 08:38:59.993100 16153 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0917 08:38:59.993130 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0917 08:38:59.994854 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0917 08:39:00.076404 16153 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0917 08:39:00.076495 16153 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0917 08:39:00.076749 16153 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0917 08:39:00.076791 16153 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0917 08:39:00.083234 16153 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0917 08:39:00.083260 16153 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0917 08:39:00.086509 16153 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0917 08:39:00.086592 16153 ssh_runner.go:362] scp helm-tiller/helm-tiller-svc.yaml --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0917 08:39:00.174851 16153 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0917 08:39:00.174944 16153 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0917 08:39:00.175368 16153 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0917 08:39:00.175430 16153 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0917 08:39:00.387881 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0917 08:39:00.391769 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0917 08:39:00.490067 16153 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0917 08:39:00.490099 16153 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0917 08:39:00.580527 16153 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0917 08:39:00.580569 16153 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0917 08:39:00.673435 16153 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0917 08:39:00.673467 16153 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0917 08:39:00.681909 16153 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0917 08:39:00.681938 16153 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0917 08:39:00.784617 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0917 08:39:00.992678 16153 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0917 08:39:00.992735 16153 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0917 08:39:00.995959 16153 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0917 08:39:00.995988 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0917 08:39:01.279999 16153 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.706361447s)
I0917 08:39:01.280042 16153 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0917 08:39:01.280298 16153 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.7067792s)
I0917 08:39:01.281318 16153 node_ready.go:35] waiting up to 6m0s for node "addons-118348" to be "Ready" ...
I0917 08:39:01.284205 16153 node_ready.go:49] node "addons-118348" has status "Ready":"True"
I0917 08:39:01.284235 16153 node_ready.go:38] duration metric: took 2.882443ms for node "addons-118348" to be "Ready" ...
I0917 08:39:01.284246 16153 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0917 08:39:01.294259 16153 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace to be "Ready" ...
I0917 08:39:01.297441 16153 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0917 08:39:01.297474 16153 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0917 08:39:01.491379 16153 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0917 08:39:01.491410 16153 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0917 08:39:01.678875 16153 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0917 08:39:01.678906 16153 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0917 08:39:01.691507 16153 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0917 08:39:01.691535 16153 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0917 08:39:01.787825 16153 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-118348" context rescaled to 1 replicas
I0917 08:39:01.878797 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0917 08:39:02.379637 16153 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0917 08:39:02.379727 16153 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0917 08:39:02.590631 16153 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0917 08:39:02.590724 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0917 08:39:02.982952 16153 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0917 08:39:02.982982 16153 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0917 08:39:03.375628 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:03.576041 16153 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0917 08:39:03.576118 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0917 08:39:03.576375 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (3.976139353s)
I0917 08:39:03.576468 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (3.900368954s)
I0917 08:39:03.576791 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (3.88359943s)
I0917 08:39:03.576883 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (3.785026119s)
I0917 08:39:03.576955 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (3.686609221s)
I0917 08:39:03.588988 16153 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0917 08:39:03.589093 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0917 08:39:03.783768 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0917 08:39:03.793536 16153 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0917 08:39:03.793618 16153 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0917 08:39:03.873440 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0917 08:39:04.184222 16153 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0917 08:39:04.184256 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0917 08:39:04.584066 16153 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0917 08:39:04.584096 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0917 08:39:05.183635 16153 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0917 08:39:05.183729 16153 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0917 08:39:05.378493 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:05.483860 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0917 08:39:05.488267 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (5.513774328s)
I0917 08:39:06.091339 16153 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0917 08:39:06.091417 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:39:06.111586 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:39:06.979746 16153 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0917 08:39:07.382452 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:07.476557 16153 addons.go:234] Setting addon gcp-auth=true in "addons-118348"
I0917 08:39:07.476740 16153 host.go:66] Checking if "addons-118348" exists ...
I0917 08:39:07.477304 16153 cli_runner.go:164] Run: docker container inspect addons-118348 --format={{.State.Status}}
I0917 08:39:07.505101 16153 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0917 08:39:07.505159 16153 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-118348
I0917 08:39:07.522879 16153 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-8091/.minikube/machines/addons-118348/id_rsa Username:docker}
I0917 08:39:08.791402 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (8.814611517s)
I0917 08:39:08.791526 16153 addons.go:475] Verifying addon ingress=true in "addons-118348"
I0917 08:39:08.794143 16153 out.go:177] * Verifying ingress addon...
I0917 08:39:08.797657 16153 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0917 08:39:08.801435 16153 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0917 08:39:08.801494 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:09.378157 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:09.882501 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:09.889642 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:10.383665 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:10.881625 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:11.377600 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:11.803330 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:11.890541 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:11.895607 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.90071276s)
I0917 08:39:11.895789 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (11.507869181s)
I0917 08:39:11.895816 16153 addons.go:475] Verifying addon metrics-server=true in "addons-118348"
I0917 08:39:11.895910 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (11.504082402s)
I0917 08:39:11.895944 16153 addons.go:475] Verifying addon registry=true in "addons-118348"
I0917 08:39:11.896030 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (11.111369791s)
I0917 08:39:11.896072 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (10.017176357s)
I0917 08:39:11.896308 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (8.11243331s)
W0917 08:39:11.896404 16153 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0917 08:39:11.896427 16153 retry.go:31] will retry after 322.153826ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0917 08:39:11.896433 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (8.02289686s)
I0917 08:39:11.897620 16153 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-118348 service yakd-dashboard -n yakd-dashboard
I0917 08:39:11.897734 16153 out.go:177] * Verifying registry addon...
I0917 08:39:11.899790 16153 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0917 08:39:11.976198 16153 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0917 08:39:11.976232 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:12.219600 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0917 08:39:12.302804 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:12.491547 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:12.801971 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:12.904268 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:13.186191 16153 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.681058063s)
I0917 08:39:13.186448 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (7.7022062s)
I0917 08:39:13.186488 16153 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-118348"
I0917 08:39:13.188467 16153 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0917 08:39:13.188469 16153 out.go:177] * Verifying csi-hostpath-driver addon...
I0917 08:39:13.189927 16153 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0917 08:39:13.190864 16153 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0917 08:39:13.191217 16153 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0917 08:39:13.191240 16153 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0917 08:39:13.197784 16153 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0917 08:39:13.197815 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:13.286180 16153 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0917 08:39:13.286207 16153 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0917 08:39:13.301069 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:13.374751 16153 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0917 08:39:13.374829 16153 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0917 08:39:13.398688 16153 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0917 08:39:13.407581 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:13.697225 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:13.876265 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:13.976651 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:14.195368 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:14.301336 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:14.374564 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:14.476158 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:14.696223 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:14.801195 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:14.805561 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.585910854s)
I0917 08:39:14.903124 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:14.989553 16153 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.590823205s)
I0917 08:39:14.990773 16153 addons.go:475] Verifying addon gcp-auth=true in "addons-118348"
I0917 08:39:14.992673 16153 out.go:177] * Verifying gcp-auth addon...
I0917 08:39:14.995175 16153 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0917 08:39:15.003474 16153 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0917 08:39:15.196047 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:15.301091 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:15.403998 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:15.695926 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:15.800975 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:15.903517 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:16.196433 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:16.300641 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:16.403886 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:16.695937 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:16.800195 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:16.801159 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:16.903787 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:17.195800 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:17.300614 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:17.404447 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:17.695753 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:17.801089 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:17.903504 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:18.240207 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:18.300778 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:18.403201 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:18.694615 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:18.800937 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:18.902955 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:19.195447 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:19.300374 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:19.301314 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:19.404212 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:19.695526 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:19.801550 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:19.903670 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:20.196313 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:20.301097 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:20.403515 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:20.695785 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:20.801697 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:20.903754 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:21.196407 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:21.300444 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:21.301884 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:21.403475 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:21.696976 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:21.801174 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:21.903760 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:22.195237 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:22.302075 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:22.403564 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:22.695762 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:22.801078 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:22.903394 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:23.195852 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:23.300985 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:23.403672 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:23.695507 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:23.799824 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:23.801331 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:23.903684 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:24.195953 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:24.301016 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:24.403280 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:24.695823 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:24.801714 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:24.903787 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:25.196103 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:25.301728 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:25.403545 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:25.696288 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:25.800494 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:25.801279 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:25.903645 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:26.195640 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:26.301106 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:26.403181 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:26.695080 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:26.801986 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:26.904869 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:27.196098 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:27.300924 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:27.403834 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:27.703389 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:27.801061 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:27.905185 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:28.195535 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:28.300471 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:28.301181 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:28.403736 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:28.695445 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:28.806196 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:28.977196 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:29.195010 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:29.300944 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:29.403058 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:29.695632 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:29.801045 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:29.903518 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:30.235969 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:30.337102 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:30.402952 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:30.695584 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:30.800195 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:30.801344 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:30.903928 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:31.196099 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:31.300553 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:31.402910 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:31.696553 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:31.801477 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:31.903989 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:32.196221 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:32.301328 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:32.404350 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:32.695867 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:32.801174 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:32.903454 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:33.196052 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:33.300720 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:33.301711 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:33.404101 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:33.696008 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:33.801185 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:33.904266 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:34.195896 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:34.300747 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:34.403215 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:34.695586 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:34.800528 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:34.902669 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:35.195847 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:35.300842 16153 pod_ready.go:103] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"False"
I0917 08:39:35.301012 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:35.403775 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:35.696024 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:35.801514 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:35.903751 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:36.195985 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:36.300801 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:36.403241 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:36.695555 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:36.800821 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:36.903255 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:37.197023 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:37.300348 16153 pod_ready.go:93] pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace has status "Ready":"True"
I0917 08:39:37.300374 16153 pod_ready.go:82] duration metric: took 36.006078672s for pod "coredns-7c65d6cfc9-25csd" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.300387 16153 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-fd7jf" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.301524 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:37.302193 16153 pod_ready.go:98] error getting pod "coredns-7c65d6cfc9-fd7jf" in "kube-system" namespace (skipping!): pods "coredns-7c65d6cfc9-fd7jf" not found
I0917 08:39:37.302215 16153 pod_ready.go:82] duration metric: took 1.819791ms for pod "coredns-7c65d6cfc9-fd7jf" in "kube-system" namespace to be "Ready" ...
E0917 08:39:37.302227 16153 pod_ready.go:67] WaitExtra: waitPodCondition: error getting pod "coredns-7c65d6cfc9-fd7jf" in "kube-system" namespace (skipping!): pods "coredns-7c65d6cfc9-fd7jf" not found
I0917 08:39:37.302236 16153 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-118348" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.306831 16153 pod_ready.go:93] pod "etcd-addons-118348" in "kube-system" namespace has status "Ready":"True"
I0917 08:39:37.306853 16153 pod_ready.go:82] duration metric: took 4.609235ms for pod "etcd-addons-118348" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.306866 16153 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-118348" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.311486 16153 pod_ready.go:93] pod "kube-apiserver-addons-118348" in "kube-system" namespace has status "Ready":"True"
I0917 08:39:37.311511 16153 pod_ready.go:82] duration metric: took 4.63615ms for pod "kube-apiserver-addons-118348" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.311524 16153 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-118348" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.316429 16153 pod_ready.go:93] pod "kube-controller-manager-addons-118348" in "kube-system" namespace has status "Ready":"True"
I0917 08:39:37.316450 16153 pod_ready.go:82] duration metric: took 4.917819ms for pod "kube-controller-manager-addons-118348" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.316459 16153 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-kbbwc" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.403933 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:37.498298 16153 pod_ready.go:93] pod "kube-proxy-kbbwc" in "kube-system" namespace has status "Ready":"True"
I0917 08:39:37.498331 16153 pod_ready.go:82] duration metric: took 181.864769ms for pod "kube-proxy-kbbwc" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.498344 16153 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-118348" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.695245 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:37.801443 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:37.898198 16153 pod_ready.go:93] pod "kube-scheduler-addons-118348" in "kube-system" namespace has status "Ready":"True"
I0917 08:39:37.898221 16153 pod_ready.go:82] duration metric: took 399.869471ms for pod "kube-scheduler-addons-118348" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.898232 16153 pod_ready.go:79] waiting up to 6m0s for pod "nvidia-device-plugin-daemonset-sghds" in "kube-system" namespace to be "Ready" ...
I0917 08:39:37.902525 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:38.301302 16153 pod_ready.go:93] pod "nvidia-device-plugin-daemonset-sghds" in "kube-system" namespace has status "Ready":"True"
I0917 08:39:38.301335 16153 pod_ready.go:82] duration metric: took 403.095412ms for pod "nvidia-device-plugin-daemonset-sghds" in "kube-system" namespace to be "Ready" ...
I0917 08:39:38.301345 16153 pod_ready.go:39] duration metric: took 37.017085561s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0917 08:39:38.301368 16153 api_server.go:52] waiting for apiserver process to appear ...
I0917 08:39:38.301428 16153 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0917 08:39:38.302262 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:38.302424 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:38.315708 16153 api_server.go:72] duration metric: took 39.275221698s to wait for apiserver process to appear ...
I0917 08:39:38.315733 16153 api_server.go:88] waiting for apiserver healthz status ...
I0917 08:39:38.315756 16153 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0917 08:39:38.319961 16153 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0917 08:39:38.320776 16153 api_server.go:141] control plane version: v1.31.1
I0917 08:39:38.320797 16153 api_server.go:131] duration metric: took 5.057934ms to wait for apiserver health ...
I0917 08:39:38.320804 16153 system_pods.go:43] waiting for kube-system pods to appear ...
I0917 08:39:38.403547 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:38.504076 16153 system_pods.go:59] 18 kube-system pods found
I0917 08:39:38.504116 16153 system_pods.go:61] "coredns-7c65d6cfc9-25csd" [74c6e7e6-faf9-4bf7-9fa9-534033b67fba] Running
I0917 08:39:38.504127 16153 system_pods.go:61] "csi-hostpath-attacher-0" [73ce30f8-48fd-4ad4-baf1-931bcb63ef19] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0917 08:39:38.504135 16153 system_pods.go:61] "csi-hostpath-resizer-0" [f17ba8bf-d93d-4035-a806-a89e1efd8207] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0917 08:39:38.504145 16153 system_pods.go:61] "csi-hostpathplugin-fdkjh" [3613a36d-0f3a-4229-9fa1-dd07229fc18e] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0917 08:39:38.504159 16153 system_pods.go:61] "etcd-addons-118348" [c46ee4c9-ef8b-4058-9b7d-68c96e840ec2] Running
I0917 08:39:38.504165 16153 system_pods.go:61] "kube-apiserver-addons-118348" [b3a3e80d-da7d-4156-8dd0-175c76c46ee1] Running
I0917 08:39:38.504171 16153 system_pods.go:61] "kube-controller-manager-addons-118348" [abd17890-f547-4618-82eb-0e928bc66c10] Running
I0917 08:39:38.504176 16153 system_pods.go:61] "kube-ingress-dns-minikube" [62aae420-556f-422c-92b2-ee34cf2cf9df] Running
I0917 08:39:38.504181 16153 system_pods.go:61] "kube-proxy-kbbwc" [0ad6532a-3039-47bf-9e87-8cb8503cb75b] Running
I0917 08:39:38.504185 16153 system_pods.go:61] "kube-scheduler-addons-118348" [86664b5d-c607-4a88-be3e-8839052425e3] Running
I0917 08:39:38.504192 16153 system_pods.go:61] "metrics-server-84c5f94fbc-9dxps" [3646ec2c-2273-4bf7-af3e-a3dfe0d91552] Running
I0917 08:39:38.504200 16153 system_pods.go:61] "nvidia-device-plugin-daemonset-sghds" [1dd15af2-9e1e-4296-99f7-992a66fc0483] Running
I0917 08:39:38.504205 16153 system_pods.go:61] "registry-66c9cd494c-t5sv4" [2f41b6f7-f293-467f-8215-b24af50ec8ba] Running
I0917 08:39:38.504213 16153 system_pods.go:61] "registry-proxy-z9ss9" [29edb9a3-341b-486a-8045-5546e8911d8c] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0917 08:39:38.504228 16153 system_pods.go:61] "snapshot-controller-56fcc65765-bdpj5" [a10b3217-9bdc-40fd-8d71-3b72e5228e60] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0917 08:39:38.504238 16153 system_pods.go:61] "snapshot-controller-56fcc65765-t4r9g" [d3c2bbc8-bdca-4d5d-bed7-90bc3f95662b] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0917 08:39:38.504247 16153 system_pods.go:61] "storage-provisioner" [7704bba1-fdaa-4b34-8549-a04eb5b45b4a] Running
I0917 08:39:38.504255 16153 system_pods.go:61] "tiller-deploy-b48cc5f79-ng9ss" [cad55614-1ecf-4037-9bff-258c6c00984a] Running
I0917 08:39:38.504266 16153 system_pods.go:74] duration metric: took 183.454802ms to wait for pod list to return data ...
I0917 08:39:38.504278 16153 default_sa.go:34] waiting for default service account to be created ...
I0917 08:39:38.697019 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:38.697288 16153 default_sa.go:45] found service account: "default"
I0917 08:39:38.697310 16153 default_sa.go:55] duration metric: took 193.024971ms for default service account to be created ...
I0917 08:39:38.697320 16153 system_pods.go:116] waiting for k8s-apps to be running ...
I0917 08:39:38.801769 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:38.903378 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 08:39:38.904118 16153 system_pods.go:86] 18 kube-system pods found
I0917 08:39:38.904139 16153 system_pods.go:89] "coredns-7c65d6cfc9-25csd" [74c6e7e6-faf9-4bf7-9fa9-534033b67fba] Running
I0917 08:39:38.904149 16153 system_pods.go:89] "csi-hostpath-attacher-0" [73ce30f8-48fd-4ad4-baf1-931bcb63ef19] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0917 08:39:38.904155 16153 system_pods.go:89] "csi-hostpath-resizer-0" [f17ba8bf-d93d-4035-a806-a89e1efd8207] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0917 08:39:38.904163 16153 system_pods.go:89] "csi-hostpathplugin-fdkjh" [3613a36d-0f3a-4229-9fa1-dd07229fc18e] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0917 08:39:38.904171 16153 system_pods.go:89] "etcd-addons-118348" [c46ee4c9-ef8b-4058-9b7d-68c96e840ec2] Running
I0917 08:39:38.904176 16153 system_pods.go:89] "kube-apiserver-addons-118348" [b3a3e80d-da7d-4156-8dd0-175c76c46ee1] Running
I0917 08:39:38.904180 16153 system_pods.go:89] "kube-controller-manager-addons-118348" [abd17890-f547-4618-82eb-0e928bc66c10] Running
I0917 08:39:38.904187 16153 system_pods.go:89] "kube-ingress-dns-minikube" [62aae420-556f-422c-92b2-ee34cf2cf9df] Running
I0917 08:39:38.904193 16153 system_pods.go:89] "kube-proxy-kbbwc" [0ad6532a-3039-47bf-9e87-8cb8503cb75b] Running
I0917 08:39:38.904199 16153 system_pods.go:89] "kube-scheduler-addons-118348" [86664b5d-c607-4a88-be3e-8839052425e3] Running
I0917 08:39:38.904203 16153 system_pods.go:89] "metrics-server-84c5f94fbc-9dxps" [3646ec2c-2273-4bf7-af3e-a3dfe0d91552] Running
I0917 08:39:38.904207 16153 system_pods.go:89] "nvidia-device-plugin-daemonset-sghds" [1dd15af2-9e1e-4296-99f7-992a66fc0483] Running
I0917 08:39:38.904210 16153 system_pods.go:89] "registry-66c9cd494c-t5sv4" [2f41b6f7-f293-467f-8215-b24af50ec8ba] Running
I0917 08:39:38.904216 16153 system_pods.go:89] "registry-proxy-z9ss9" [29edb9a3-341b-486a-8045-5546e8911d8c] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0917 08:39:38.904225 16153 system_pods.go:89] "snapshot-controller-56fcc65765-bdpj5" [a10b3217-9bdc-40fd-8d71-3b72e5228e60] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0917 08:39:38.904232 16153 system_pods.go:89] "snapshot-controller-56fcc65765-t4r9g" [d3c2bbc8-bdca-4d5d-bed7-90bc3f95662b] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0917 08:39:38.904238 16153 system_pods.go:89] "storage-provisioner" [7704bba1-fdaa-4b34-8549-a04eb5b45b4a] Running
I0917 08:39:38.904242 16153 system_pods.go:89] "tiller-deploy-b48cc5f79-ng9ss" [cad55614-1ecf-4037-9bff-258c6c00984a] Running
I0917 08:39:38.904248 16153 system_pods.go:126] duration metric: took 206.922779ms to wait for k8s-apps to be running ...
I0917 08:39:38.904258 16153 system_svc.go:44] waiting for kubelet service to be running ....
I0917 08:39:38.904339 16153 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0917 08:39:38.917345 16153 system_svc.go:56] duration metric: took 13.075694ms WaitForService to wait for kubelet
I0917 08:39:38.917378 16153 kubeadm.go:582] duration metric: took 39.876897298s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0917 08:39:38.917401 16153 node_conditions.go:102] verifying NodePressure condition ...
I0917 08:39:39.098073 16153 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0917 08:39:39.098105 16153 node_conditions.go:123] node cpu capacity is 8
I0917 08:39:39.098123 16153 node_conditions.go:105] duration metric: took 180.71542ms to run NodePressure ...
I0917 08:39:39.098138 16153 start.go:241] waiting for startup goroutines ...
I0917 08:39:39.195791 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:39.302791 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:39.403872 16153 kapi.go:107] duration metric: took 27.504080322s to wait for kubernetes.io/minikube-addons=registry ...
I0917 08:39:39.695436 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:39.802393 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:40.195103 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:40.301401 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:40.730516 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:40.831208 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:41.195300 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:41.302491 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:41.696171 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:41.829296 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:42.195488 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:42.301552 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:42.697259 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:42.801660 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:43.195280 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:43.301546 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:43.695979 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:43.801682 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:44.196023 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:44.301946 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:44.696613 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:44.802429 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:45.195525 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:45.301467 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:45.695966 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:45.802561 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:46.196104 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:46.301945 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:46.696438 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:46.825571 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:47.195080 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:47.301720 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:47.696178 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:47.802045 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:48.195035 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:48.301631 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:48.694986 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:48.801976 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:49.195128 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:49.302646 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:49.696354 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:49.802446 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:50.196238 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:50.301860 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:50.696185 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:50.802423 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:51.195880 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:51.302498 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:51.696568 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:51.806601 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:52.195437 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:52.302602 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:52.695992 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:52.802752 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:53.195662 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:53.301968 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:53.695313 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:53.802173 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:54.203372 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:54.301608 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:54.695195 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:54.801893 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:55.195571 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:55.302370 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:55.695188 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:55.802351 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:56.195986 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:56.301661 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:56.694791 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 08:39:56.801629 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:57.195174 16153 kapi.go:107] duration metric: took 44.004310731s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0917 08:39:57.301598 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:57.802687 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:58.301850 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:58.801119 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:59.301511 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:39:59.802119 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:00.301838 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:00.801943 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:01.301426 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:01.802312 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:02.301115 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:02.801956 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:03.301756 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:03.802655 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:04.301452 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:04.802272 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:05.301862 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:05.801855 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:06.301478 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:06.801733 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:07.302056 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:07.801947 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:08.301856 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:08.800916 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:09.301262 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:09.802065 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:10.301019 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:10.802132 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:11.301777 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:11.802855 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:12.301795 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:12.802699 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:13.302893 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:13.801796 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:14.301200 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:14.802589 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:15.302018 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:15.802881 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:16.302145 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:16.801783 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:17.301850 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:17.802517 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:18.302340 16153 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 08:40:18.800977 16153 kapi.go:107] duration metric: took 1m10.003320251s to wait for app.kubernetes.io/name=ingress-nginx ...
I0917 08:40:37.000995 16153 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0917 08:40:37.001020 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:37.498839 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:37.999234 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:38.497864 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:38.999407 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:39.498109 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:40.000492 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:40.498386 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:40.998451 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:41.498255 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:41.998602 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:42.498448 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:42.998865 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:43.498896 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:44.024262 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:44.498031 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:45.001240 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:45.497726 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:45.998983 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:46.498886 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:47.000667 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:47.498370 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:47.998495 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:48.498487 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:48.998690 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:49.499086 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:49.998549 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:50.497985 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:51.001190 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:51.498227 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:52.000369 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:52.498649 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:52.998951 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:53.498953 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:54.000937 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:54.499007 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:54.999326 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:55.497932 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:55.999432 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:56.498395 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:56.998815 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:57.498643 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:57.999117 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:58.498698 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:58.998740 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:59.498680 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:40:59.999109 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:00.498965 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:00.999078 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:01.497822 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:02.001306 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:02.498441 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:03.000381 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:03.498158 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:03.998500 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:04.498506 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:05.000068 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:05.497808 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:06.000568 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:06.498681 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:06.998371 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:07.498473 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:07.998732 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:08.498790 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:09.000611 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:09.498584 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:10.000713 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:10.498468 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:11.000596 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:11.498422 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:11.998431 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:12.498404 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:12.998461 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:13.498426 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:14.001236 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:14.498302 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:14.998380 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:15.498510 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:15.998709 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:16.498753 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:17.000790 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:17.498782 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:18.000746 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:18.499228 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:19.000539 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:19.498787 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:19.999307 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:20.498234 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:21.000895 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:21.498207 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:21.998612 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:22.498774 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:23.000562 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:23.498424 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:23.998669 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:24.498853 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:24.999126 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:25.498070 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:26.001037 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:26.498709 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:27.001781 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:27.498891 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:27.999036 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:28.498858 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:28.998778 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:29.498500 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:29.998060 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:30.497718 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:30.998714 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:31.498270 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:32.000792 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:32.498578 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:33.000358 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:33.498107 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:34.000251 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:34.498255 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:34.998231 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:35.498260 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:35.998631 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:36.498345 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:36.998180 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:37.498406 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:37.999412 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:38.497992 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:38.998965 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:39.497768 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:40.000952 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:40.498410 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:41.000926 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:41.499104 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:42.001070 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:42.497835 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:42.999533 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:43.498244 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:43.998439 16153 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 08:41:44.498642 16153 kapi.go:107] duration metric: took 2m29.50346527s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0917 08:41:44.500599 16153 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-118348 cluster.
I0917 08:41:44.502087 16153 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0917 08:41:44.503315 16153 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0917 08:41:44.504843 16153 out.go:177] * Enabled addons: storage-provisioner, nvidia-device-plugin, ingress-dns, cloud-spanner, default-storageclass, storage-provisioner-rancher, volcano, metrics-server, helm-tiller, inspektor-gadget, yakd, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0917 08:41:44.506400 16153 addons.go:510] duration metric: took 2m45.465894838s for enable addons: enabled=[storage-provisioner nvidia-device-plugin ingress-dns cloud-spanner default-storageclass storage-provisioner-rancher volcano metrics-server helm-tiller inspektor-gadget yakd volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0917 08:41:44.506453 16153 start.go:246] waiting for cluster config update ...
I0917 08:41:44.506486 16153 start.go:255] writing updated cluster config ...
I0917 08:41:44.506795 16153 ssh_runner.go:195] Run: rm -f paused
I0917 08:41:44.554434 16153 start.go:600] kubectl: 1.31.1, cluster: 1.31.1 (minor skew: 0)
I0917 08:41:44.556262 16153 out.go:177] * Done! kubectl is now configured to use "addons-118348" cluster and "default" namespace by default
==> Docker <==
Sep 17 08:51:10 addons-118348 dockerd[1336]: time="2024-09-17T08:51:10.973742658Z" level=info msg="ignoring event" container=cc3fbeb2435cc6c4788b18c17bae5c68840eccb43158fabff407fe3a0b337d01 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:11 addons-118348 dockerd[1336]: time="2024-09-17T08:51:11.112830939Z" level=info msg="ignoring event" container=5bea26d60bddb65cf782a4e101a6ce375d64ca9756ed19dfefcc38a3d05423bc module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:11 addons-118348 dockerd[1336]: time="2024-09-17T08:51:11.199655217Z" level=info msg="ignoring event" container=ce2f10f9d98f6c27aedb86605306b03715140f3e48c59d1310e41747b4a7dc69 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:11 addons-118348 dockerd[1336]: time="2024-09-17T08:51:11.274295423Z" level=info msg="ignoring event" container=288210703f58c42010f08593054289bf2d51e7672da207c991e080596682d56c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:11 addons-118348 dockerd[1336]: time="2024-09-17T08:51:11.274597037Z" level=info msg="ignoring event" container=f92be85642967b20987efd3b92788cbde05ac27f455ca9dd4b2f987761334625 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:11 addons-118348 cri-dockerd[1601]: time="2024-09-17T08:51:11Z" level=info msg="Stop pulling image docker.io/kicbase/echo-server:1.0: Status: Downloaded newer image for kicbase/echo-server:1.0"
Sep 17 08:51:15 addons-118348 dockerd[1336]: time="2024-09-17T08:51:15.525865827Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=229912276e6a1ada3f17bfdb913a4925e0c346a45c48d462fc006bfe59ce64c0
Sep 17 08:51:15 addons-118348 dockerd[1336]: time="2024-09-17T08:51:15.569849678Z" level=info msg="ignoring event" container=229912276e6a1ada3f17bfdb913a4925e0c346a45c48d462fc006bfe59ce64c0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:15 addons-118348 dockerd[1336]: time="2024-09-17T08:51:15.715058693Z" level=info msg="ignoring event" container=ef9a28f1d568c6e8e2b29ccf81df016e1930071eb0423a09c7099c8989b59077 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:16 addons-118348 dockerd[1336]: time="2024-09-17T08:51:16.404089885Z" level=info msg="Container failed to exit within 30s of signal 15 - using the force" container=34ebf1568823a9abdb506aad920db2aa7508f251a8feebd84bcef40f9f0a176c
Sep 17 08:51:16 addons-118348 dockerd[1336]: time="2024-09-17T08:51:16.425897851Z" level=info msg="ignoring event" container=34ebf1568823a9abdb506aad920db2aa7508f251a8feebd84bcef40f9f0a176c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:16 addons-118348 dockerd[1336]: time="2024-09-17T08:51:16.533659046Z" level=info msg="ignoring event" container=17107eab7e2c053d5f508267e4216e7b2f2748ed46998fc843aa6ce26e9e6cc0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:16 addons-118348 dockerd[1336]: time="2024-09-17T08:51:16.586035038Z" level=info msg="ignoring event" container=383440bc3e922d5393564a2b25f52b2ee4c79a29d7fee1ffd176edbd3e181e8c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:16 addons-118348 dockerd[1336]: time="2024-09-17T08:51:16.694998563Z" level=info msg="ignoring event" container=737e486e424f2851ff8113b42d56ee69ac9a3096a57d1f86af8ac8490993a4ce module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:17 addons-118348 cri-dockerd[1601]: time="2024-09-17T08:51:17Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/7e79906ca769bfe62340e57232ca756180eaf5fd86fec639addce8332a05a29b/resolv.conf as [nameserver 10.96.0.10 search headlamp.svc.cluster.local svc.cluster.local cluster.local us-east4-a.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 17 08:51:18 addons-118348 dockerd[1336]: time="2024-09-17T08:51:18.048764263Z" level=warning msg="reference for unknown type: " digest="sha256:8825bb13459c64dcf9503d836b94b49c97dc831aff7c325a6eed68961388cf9c" remote="ghcr.io/headlamp-k8s/headlamp@sha256:8825bb13459c64dcf9503d836b94b49c97dc831aff7c325a6eed68961388cf9c"
Sep 17 08:51:20 addons-118348 cri-dockerd[1601]: time="2024-09-17T08:51:20Z" level=info msg="Stop pulling image ghcr.io/headlamp-k8s/headlamp:v0.25.1@sha256:8825bb13459c64dcf9503d836b94b49c97dc831aff7c325a6eed68961388cf9c: Status: Downloaded newer image for ghcr.io/headlamp-k8s/headlamp@sha256:8825bb13459c64dcf9503d836b94b49c97dc831aff7c325a6eed68961388cf9c"
Sep 17 08:51:20 addons-118348 dockerd[1336]: time="2024-09-17T08:51:20.262319605Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 17 08:51:20 addons-118348 dockerd[1336]: time="2024-09-17T08:51:20.360786649Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 17 08:51:37 addons-118348 dockerd[1336]: time="2024-09-17T08:51:37.705734164Z" level=info msg="ignoring event" container=af3707355027c5e154a2c2cbc6a1009f448196a1e252c3b0617fe33f494b9547 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:38 addons-118348 dockerd[1336]: time="2024-09-17T08:51:38.180066631Z" level=info msg="ignoring event" container=2142006050fe3d6472faea03ddf0ca8a2c6684055d4a029a2fe32467d6e90142 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:38 addons-118348 dockerd[1336]: time="2024-09-17T08:51:38.244368670Z" level=info msg="ignoring event" container=59629b206e29a2b335b54b2083abd3fdb2030d9f8f50c8e75301fce625a96c34 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:38 addons-118348 dockerd[1336]: time="2024-09-17T08:51:38.310873870Z" level=info msg="ignoring event" container=ce7007c1fbe071bf8be031cab9076d8b43d8ae3b6fb04886de006ba667f7a4e0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 08:51:38 addons-118348 cri-dockerd[1601]: time="2024-09-17T08:51:38Z" level=info msg="Failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod \"registry-proxy-z9ss9_kube-system\": unexpected command output nsenter: cannot open /proc/4131/ns/net: No such file or directory\n with error: exit status 1"
Sep 17 08:51:38 addons-118348 dockerd[1336]: time="2024-09-17T08:51:38.395824559Z" level=info msg="ignoring event" container=4f57112a8c3563bd5083401d9c67511eb4e68785c64dc9aeb47adef7c5aec4af module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
4b522f18d071e ghcr.io/headlamp-k8s/headlamp@sha256:8825bb13459c64dcf9503d836b94b49c97dc831aff7c325a6eed68961388cf9c 19 seconds ago Running headlamp 0 7e79906ca769b headlamp-7b5c95b59d-7tpl7
d35cd61b18ff7 kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 28 seconds ago Running hello-world-app 0 0a4f4694eba34 hello-world-app-55bf9c44b4-m6vxk
4f1f9e1fad01f nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf 38 seconds ago Running nginx 0 52e40a84304f3 nginx
b1ba8e0c3a913 gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 961018bc6d86e gcp-auth-89d5ffd79-5cd98
fde3cb4cbfcda ce263a8653f9c 11 minutes ago Exited patch 1 941c928e6fdd0 ingress-nginx-admission-patch-zmqg2
175f41e84f172 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 0cdd60007cdb1 ingress-nginx-admission-create-qdddq
59629b206e29a gcr.io/k8s-minikube/kube-registry-proxy@sha256:b3fa0b2df8737fdb85ad5918a7e2652527463e357afff83a5e5bb966bcedc367 12 minutes ago Exited registry-proxy 0 4f57112a8c356 registry-proxy-z9ss9
d9dc593c31cfe 6e38f40d628db 12 minutes ago Running storage-provisioner 0 5aef938fa147f storage-provisioner
b282494940988 c69fa2e9cbf5f 12 minutes ago Running coredns 0 1571a212d6e0e coredns-7c65d6cfc9-25csd
bbf2f71545987 60c005f310ff3 12 minutes ago Running kube-proxy 0 0ca66b293eb23 kube-proxy-kbbwc
632fef3907db7 175ffd71cce3d 12 minutes ago Running kube-controller-manager 0 b8c068cbe92d4 kube-controller-manager-addons-118348
3b86720cba6d5 6bab7719df100 12 minutes ago Running kube-apiserver 0 44b0deecee93b kube-apiserver-addons-118348
a87794d0bfcfc 9aa1fad941575 12 minutes ago Running kube-scheduler 0 eb92559371368 kube-scheduler-addons-118348
f68d38443e29f 2e96e5913fc06 12 minutes ago Running etcd 0 dbdce7e538d98 etcd-addons-118348
==> coredns [b28249494098] <==
[INFO] 10.244.0.22:33250 - 31260 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003903857s
[INFO] 10.244.0.22:34780 - 39607 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003875029s
[INFO] 10.244.0.22:44235 - 33312 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000078672s
[INFO] 10.244.0.22:33250 - 51374 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000061981s
[INFO] 10.244.0.22:34780 - 21876 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000044802s
[INFO] 10.244.0.22:58125 - 47928 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.00211762s
[INFO] 10.244.0.22:45801 - 16343 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.007450403s
[INFO] 10.244.0.22:43230 - 5612 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.002275494s
[INFO] 10.244.0.22:56104 - 39511 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.002220231s
[INFO] 10.244.0.22:43861 - 33385 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.002209809s
[INFO] 10.244.0.22:45801 - 19158 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005381891s
[INFO] 10.244.0.22:58125 - 34386 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005537849s
[INFO] 10.244.0.22:43861 - 1243 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005358587s
[INFO] 10.244.0.22:56104 - 49318 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005503151s
[INFO] 10.244.0.22:43230 - 17559 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005489859s
[INFO] 10.244.0.22:56104 - 60953 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004506226s
[INFO] 10.244.0.22:45801 - 22069 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00484831s
[INFO] 10.244.0.22:58125 - 49347 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00251105s
[INFO] 10.244.0.22:56104 - 16363 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000047949s
[INFO] 10.244.0.22:43230 - 29112 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.002655104s
[INFO] 10.244.0.22:43861 - 38546 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004782989s
[INFO] 10.244.0.22:43861 - 46633 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000075892s
[INFO] 10.244.0.22:58125 - 52218 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000073019s
[INFO] 10.244.0.22:45801 - 20959 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000091201s
[INFO] 10.244.0.22:43230 - 53652 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000088634s
==> describe nodes <==
Name: addons-118348
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-118348
kubernetes.io/os=linux
minikube.k8s.io/commit=9256ba43b41ea130fa48757ddb8d93db00574f61
minikube.k8s.io/name=addons-118348
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_17T08_38_54_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-118348
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Tue, 17 Sep 2024 08:38:51 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-118348
AcquireTime: <unset>
RenewTime: Tue, 17 Sep 2024 08:51:37 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Tue, 17 Sep 2024 08:51:29 +0000 Tue, 17 Sep 2024 08:38:49 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Tue, 17 Sep 2024 08:51:29 +0000 Tue, 17 Sep 2024 08:38:49 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Tue, 17 Sep 2024 08:51:29 +0000 Tue, 17 Sep 2024 08:38:49 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Tue, 17 Sep 2024 08:51:29 +0000 Tue, 17 Sep 2024 08:38:51 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-118348
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
System Info:
Machine ID: 014fe64c43e8487e9a13747dfcac2bbe
System UUID: 6e85f00e-e142-4af7-9baf-424820c40175
Boot ID: 56c7860f-74df-456c-8d25-e851e670c43e
Kernel Version: 5.15.0-1069-gcp
OS Image: Ubuntu 22.04.4 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.2.1
Kubelet Version: v1.31.1
Kube-Proxy Version: v1.31.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (12 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m15s
default hello-world-app-55bf9c44b4-m6vxk 0 (0%) 0 (0%) 0 (0%) 0 (0%) 30s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 39s
gcp-auth gcp-auth-89d5ffd79-5cd98 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
headlamp headlamp-7b5c95b59d-7tpl7 0 (0%) 0 (0%) 0 (0%) 0 (0%) 22s
kube-system coredns-7c65d6cfc9-25csd 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-118348 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-118348 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-118348 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-kbbwc 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-118348 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeHasSufficientMemory 12m (x7 over 12m) kubelet Node addons-118348 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m (x6 over 12m) kubelet Node addons-118348 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m (x6 over 12m) kubelet Node addons-118348 status is now: NodeHasSufficientPID
Normal Starting 12m kubelet Starting kubelet.
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-118348 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-118348 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-118348 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-118348 event: Registered Node addons-118348 in Controller
==> dmesg <==
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff 0a 37 71 78 45 4a 08 06
[ +1.313513] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 1a c5 48 0e 64 c7 08 06
[ +5.940819] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 1a 57 66 18 23 35 08 06
[ +0.240878] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 72 f2 ed 40 5d 18 08 06
[ +0.059493] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 46 25 f0 cc 1d 5d 08 06
[Sep17 08:40] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 9e 62 9c 3d 95 97 08 06
[Sep17 08:41] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 96 ec f4 5f 7b e9 08 06
[ +0.074008] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 02 e2 37 f0 e8 fc 08 06
[ +24.575841] IPv4: martian source 10.244.0.1 from 10.244.0.26, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff e6 2c 62 9b b4 56 08 06
[ +0.000483] IPv4: martian source 10.244.0.26 from 10.244.0.2, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff f2 64 4b 88 7b 14 08 06
[Sep17 08:50] IPv4: martian source 10.244.0.1 from 10.244.0.32, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 5e 66 c8 65 f7 ef 08 06
[Sep17 08:51] IPv4: martian source 10.244.0.36 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 9e 62 9c 3d 95 97 08 06
[ +0.674983] IPv4: martian source 10.244.0.22 from 10.244.0.2, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff f2 64 4b 88 7b 14 08 06
==> etcd [f68d38443e29] <==
{"level":"info","ts":"2024-09-17T08:38:49.298199Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-17T08:38:49.298217Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-09-17T08:38:49.298229Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-17T08:38:49.299275Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-118348 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-09-17T08:38:49.299275Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-17T08:38:49.299281Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-17T08:38:49.299305Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-17T08:38:49.299577Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-17T08:38:49.299598Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-17T08:38:49.299976Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-17T08:38:49.300085Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-17T08:38:49.300105Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-17T08:38:49.300392Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-17T08:38:49.300476Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-17T08:38:49.302401Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-17T08:38:49.302795Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"warn","ts":"2024-09-17T08:39:18.237066Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"122.49774ms","expected-duration":"100ms","prefix":"","request":"header:<ID:8128031952476417337 > lease_revoke:<id:70cc91ff2261a8bc>","response":"size:29"}
{"level":"warn","ts":"2024-09-17T08:39:38.299554Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"120.002915ms","expected-duration":"100ms","prefix":"","request":"header:<ID:8128031952476417815 > lease_revoke:<id:70cc91ff2261ad8f>","response":"size:29"}
{"level":"info","ts":"2024-09-17T08:39:38.299645Z","caller":"traceutil/trace.go:171","msg":"trace[918659701] linearizableReadLoop","detail":"{readStateIndex:1110; appliedIndex:1109; }","duration":"106.639758ms","start":"2024-09-17T08:39:38.192993Z","end":"2024-09-17T08:39:38.299632Z","steps":["trace[918659701] 'read index received' (duration: 24.71µs)","trace[918659701] 'applied index is now lower than readState.Index' (duration: 106.61398ms)"],"step_count":2}
{"level":"warn","ts":"2024-09-17T08:39:38.299760Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"106.733359ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-17T08:39:38.299797Z","caller":"traceutil/trace.go:171","msg":"trace[785720166] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:1089; }","duration":"106.799404ms","start":"2024-09-17T08:39:38.192982Z","end":"2024-09-17T08:39:38.299782Z","steps":["trace[785720166] 'agreement among raft nodes before linearized reading' (duration: 106.712141ms)"],"step_count":1}
{"level":"info","ts":"2024-09-17T08:48:49.916136Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1906}
{"level":"info","ts":"2024-09-17T08:48:49.940539Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1906,"took":"23.890243ms","hash":3063729370,"current-db-size-bytes":9035776,"current-db-size":"9.0 MB","current-db-size-in-use-bytes":4993024,"current-db-size-in-use":"5.0 MB"}
{"level":"info","ts":"2024-09-17T08:48:49.940580Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":3063729370,"revision":1906,"compact-revision":-1}
{"level":"info","ts":"2024-09-17T08:51:20.531262Z","caller":"traceutil/trace.go:171","msg":"trace[1068536447] transaction","detail":"{read_only:false; response_revision:3101; number_of_response:1; }","duration":"110.306434ms","start":"2024-09-17T08:51:20.420936Z","end":"2024-09-17T08:51:20.531242Z","steps":["trace[1068536447] 'process raft request' (duration: 54.119236ms)","trace[1068536447] 'compare' (duration: 56.106033ms)"],"step_count":2}
==> gcp-auth [b1ba8e0c3a91] <==
2024/09/17 08:42:24 Ready to write response ...
2024/09/17 08:50:35 Ready to marshal response ...
2024/09/17 08:50:35 Ready to write response ...
2024/09/17 08:50:37 Ready to marshal response ...
2024/09/17 08:50:37 Ready to write response ...
2024/09/17 08:50:37 Ready to marshal response ...
2024/09/17 08:50:37 Ready to write response ...
2024/09/17 08:50:37 Ready to marshal response ...
2024/09/17 08:50:37 Ready to write response ...
2024/09/17 08:50:38 Ready to marshal response ...
2024/09/17 08:50:38 Ready to write response ...
2024/09/17 08:50:45 Ready to marshal response ...
2024/09/17 08:50:45 Ready to write response ...
2024/09/17 08:50:55 Ready to marshal response ...
2024/09/17 08:50:55 Ready to write response ...
2024/09/17 08:51:00 Ready to marshal response ...
2024/09/17 08:51:00 Ready to write response ...
2024/09/17 08:51:09 Ready to marshal response ...
2024/09/17 08:51:09 Ready to write response ...
2024/09/17 08:51:17 Ready to marshal response ...
2024/09/17 08:51:17 Ready to write response ...
2024/09/17 08:51:17 Ready to marshal response ...
2024/09/17 08:51:17 Ready to write response ...
2024/09/17 08:51:17 Ready to marshal response ...
2024/09/17 08:51:17 Ready to write response ...
==> kernel <==
08:51:39 up 34 min, 0 users, load average: 0.76, 0.67, 0.68
Linux addons-118348 5.15.0-1069-gcp #77~20.04.1-Ubuntu SMP Sun Sep 1 19:39:16 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.4 LTS"
==> kube-apiserver [3b86720cba6d] <==
W0917 08:42:16.297723 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0917 08:42:16.398666 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0917 08:42:16.807001 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0917 08:50:31.855096 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0917 08:50:32.868534 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0917 08:50:33.218096 1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
I0917 08:50:43.797133 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0917 08:51:00.184194 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0917 08:51:00.387876 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.101.47.223"}
E0917 08:51:01.782150 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
I0917 08:51:09.866188 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.108.181.132"}
I0917 08:51:10.513082 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0917 08:51:10.513130 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0917 08:51:10.584958 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0917 08:51:10.585002 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0917 08:51:10.588373 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0917 08:51:10.588416 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0917 08:51:10.602760 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0917 08:51:10.602804 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0917 08:51:10.779402 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0917 08:51:10.779456 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0917 08:51:11.585787 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W0917 08:51:11.780434 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
W0917 08:51:11.790445 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
I0917 08:51:17.204148 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.103.251.186"}
==> kube-controller-manager [632fef3907db] <==
W0917 08:51:26.280028 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 08:51:26.280073 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 08:51:28.128424 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 08:51:28.128460 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0917 08:51:28.746043 1 shared_informer.go:313] Waiting for caches to sync for resource quota
I0917 08:51:28.746084 1 shared_informer.go:320] Caches are synced for resource quota
W0917 08:51:28.779384 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 08:51:28.779428 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0917 08:51:28.977484 1 shared_informer.go:313] Waiting for caches to sync for garbage collector
I0917 08:51:28.977526 1 shared_informer.go:320] Caches are synced for garbage collector
I0917 08:51:29.312708 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-118348"
W0917 08:51:31.932154 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 08:51:31.932193 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 08:51:33.221823 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 08:51:33.221861 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0917 08:51:33.780766 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="local-path-storage"
W0917 08:51:34.312702 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 08:51:34.312740 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 08:51:36.655037 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 08:51:36.655074 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 08:51:36.921510 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 08:51:36.921556 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 08:51:37.569250 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 08:51:37.569291 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0917 08:51:38.134158 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="7.075µs"
==> kube-proxy [bbf2f7154598] <==
I0917 08:39:02.680177 1 server_linux.go:66] "Using iptables proxy"
I0917 08:39:03.279821 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0917 08:39:03.279899 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0917 08:39:03.477044 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0917 08:39:03.477105 1 server_linux.go:169] "Using iptables Proxier"
I0917 08:39:03.482215 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0917 08:39:03.482630 1 server.go:483] "Version info" version="v1.31.1"
I0917 08:39:03.482657 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0917 08:39:03.578163 1 config.go:199] "Starting service config controller"
I0917 08:39:03.578193 1 shared_informer.go:313] Waiting for caches to sync for service config
I0917 08:39:03.578217 1 config.go:105] "Starting endpoint slice config controller"
I0917 08:39:03.578221 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0917 08:39:03.578694 1 config.go:328] "Starting node config controller"
I0917 08:39:03.578703 1 shared_informer.go:313] Waiting for caches to sync for node config
I0917 08:39:03.679344 1 shared_informer.go:320] Caches are synced for node config
I0917 08:39:03.679380 1 shared_informer.go:320] Caches are synced for service config
I0917 08:39:03.679409 1 shared_informer.go:320] Caches are synced for endpoint slice config
==> kube-scheduler [a87794d0bfcf] <==
W0917 08:38:51.482118 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0917 08:38:51.482134 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 08:38:51.482149 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0917 08:38:51.482169 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 08:38:51.482201 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0917 08:38:51.482217 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 08:38:51.482245 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0917 08:38:51.482269 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0917 08:38:51.482336 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0917 08:38:51.482360 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0917 08:38:51.482553 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0917 08:38:51.482579 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 08:38:52.293377 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0917 08:38:52.293413 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 08:38:52.377731 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0917 08:38:52.377771 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0917 08:38:52.489375 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0917 08:38:52.489412 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0917 08:38:52.527687 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0917 08:38:52.527732 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0917 08:38:52.551119 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0917 08:38:52.551166 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0917 08:38:52.564769 1 reflector.go:561] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0917 08:38:52.564813 1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
I0917 08:38:54.678654 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 17 08:51:17 addons-118348 kubelet[2448]: I0917 08:51:17.806918 2448 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c49fa658-74f7-46a0-9b08-17722aeffa19" path="/var/lib/kubelet/pods/c49fa658-74f7-46a0-9b08-17722aeffa19/volumes"
Sep 17 08:51:20 addons-118348 kubelet[2448]: E0917 08:51:20.361283 2448 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = Error response from daemon: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed" image="gcr.io/k8s-minikube/busybox:latest"
Sep 17 08:51:20 addons-118348 kubelet[2448]: E0917 08:51:20.361468 2448 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-test,Image:gcr.io/k8s-minikube/busybox,Command:[],Args:[sh -c wget --spider -S http://registry.kube-system.svc.cluster.local],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:GOOGLE_APPLICATION_CREDENTIALS,Value:/google-app-creds.json,ValueFrom:nil,},EnvVar{Name:PROJECT_ID,Value:this_is_fake,ValueFrom:nil,},EnvVar{Name:GCP_PROJECT,Value:this_is_fake,ValueFrom:nil,},EnvVar{Name:GCLOUD_PROJECT,Value:this_is_fake,ValueFrom:nil,},EnvVar{Name:GOOGLE_CLOUD_PROJECT,Value:this_is_fake,ValueFrom:nil,},EnvVar{Name:CLOUDSDK_CORE_PROJECT,Value:this_is_fake,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mnhdj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,Su
bPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:gcp-creds,ReadOnly:true,MountPath:/google-app-creds.json,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:true,StdinOnce:true,TTY:true,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod registry-test_default(1fb4a1fe-7975-4387-bbb2-4911ca88db0b): ErrImagePull: Error response from daemon: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed" logger="UnhandledError"
Sep 17 08:51:20 addons-118348 kubelet[2448]: E0917 08:51:20.362658 2448 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ErrImagePull: \"Error response from daemon: Head \\\"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\\\": unauthorized: authentication failed\"" pod="default/registry-test" podUID="1fb4a1fe-7975-4387-bbb2-4911ca88db0b"
Sep 17 08:51:20 addons-118348 kubelet[2448]: E0917 08:51:20.799882 2448 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="b26b0151-c2f7-46cb-a21c-97d6bc6db827"
Sep 17 08:51:21 addons-118348 kubelet[2448]: I0917 08:51:21.215714 2448 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="headlamp/headlamp-7b5c95b59d-7tpl7" podStartSLOduration=2.023894749 podStartE2EDuration="4.215695131s" podCreationTimestamp="2024-09-17 08:51:17 +0000 UTC" firstStartedPulling="2024-09-17 08:51:18.019028834 +0000 UTC m=+744.427559713" lastFinishedPulling="2024-09-17 08:51:20.210829227 +0000 UTC m=+746.619360095" observedRunningTime="2024-09-17 08:51:21.215565487 +0000 UTC m=+747.624096393" watchObservedRunningTime="2024-09-17 08:51:21.215695131 +0000 UTC m=+747.624226013"
Sep 17 08:51:31 addons-118348 kubelet[2448]: E0917 08:51:31.800507 2448 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="b26b0151-c2f7-46cb-a21c-97d6bc6db827"
Sep 17 08:51:34 addons-118348 kubelet[2448]: E0917 08:51:34.800580 2448 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox\\\"\"" pod="default/registry-test" podUID="1fb4a1fe-7975-4387-bbb2-4911ca88db0b"
Sep 17 08:51:35 addons-118348 kubelet[2448]: I0917 08:51:35.799348 2448 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="kube-system/registry-proxy-z9ss9" secret="" err="secret \"gcp-auth\" not found"
Sep 17 08:51:37 addons-118348 kubelet[2448]: I0917 08:51:37.891514 2448 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnhdj\" (UniqueName: \"kubernetes.io/projected/1fb4a1fe-7975-4387-bbb2-4911ca88db0b-kube-api-access-mnhdj\") pod \"1fb4a1fe-7975-4387-bbb2-4911ca88db0b\" (UID: \"1fb4a1fe-7975-4387-bbb2-4911ca88db0b\") "
Sep 17 08:51:37 addons-118348 kubelet[2448]: I0917 08:51:37.891584 2448 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/1fb4a1fe-7975-4387-bbb2-4911ca88db0b-gcp-creds\") pod \"1fb4a1fe-7975-4387-bbb2-4911ca88db0b\" (UID: \"1fb4a1fe-7975-4387-bbb2-4911ca88db0b\") "
Sep 17 08:51:37 addons-118348 kubelet[2448]: I0917 08:51:37.891670 2448 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1fb4a1fe-7975-4387-bbb2-4911ca88db0b-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "1fb4a1fe-7975-4387-bbb2-4911ca88db0b" (UID: "1fb4a1fe-7975-4387-bbb2-4911ca88db0b"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Sep 17 08:51:37 addons-118348 kubelet[2448]: I0917 08:51:37.893615 2448 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb4a1fe-7975-4387-bbb2-4911ca88db0b-kube-api-access-mnhdj" (OuterVolumeSpecName: "kube-api-access-mnhdj") pod "1fb4a1fe-7975-4387-bbb2-4911ca88db0b" (UID: "1fb4a1fe-7975-4387-bbb2-4911ca88db0b"). InnerVolumeSpecName "kube-api-access-mnhdj". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 17 08:51:37 addons-118348 kubelet[2448]: I0917 08:51:37.991819 2448 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/1fb4a1fe-7975-4387-bbb2-4911ca88db0b-gcp-creds\") on node \"addons-118348\" DevicePath \"\""
Sep 17 08:51:37 addons-118348 kubelet[2448]: I0917 08:51:37.991852 2448 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-mnhdj\" (UniqueName: \"kubernetes.io/projected/1fb4a1fe-7975-4387-bbb2-4911ca88db0b-kube-api-access-mnhdj\") on node \"addons-118348\" DevicePath \"\""
Sep 17 08:51:38 addons-118348 kubelet[2448]: I0917 08:51:38.397153 2448 scope.go:117] "RemoveContainer" containerID="2142006050fe3d6472faea03ddf0ca8a2c6684055d4a029a2fe32467d6e90142"
Sep 17 08:51:38 addons-118348 kubelet[2448]: I0917 08:51:38.414580 2448 scope.go:117] "RemoveContainer" containerID="2142006050fe3d6472faea03ddf0ca8a2c6684055d4a029a2fe32467d6e90142"
Sep 17 08:51:38 addons-118348 kubelet[2448]: E0917 08:51:38.415407 2448 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 2142006050fe3d6472faea03ddf0ca8a2c6684055d4a029a2fe32467d6e90142" containerID="2142006050fe3d6472faea03ddf0ca8a2c6684055d4a029a2fe32467d6e90142"
Sep 17 08:51:38 addons-118348 kubelet[2448]: I0917 08:51:38.415448 2448 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"2142006050fe3d6472faea03ddf0ca8a2c6684055d4a029a2fe32467d6e90142"} err="failed to get container status \"2142006050fe3d6472faea03ddf0ca8a2c6684055d4a029a2fe32467d6e90142\": rpc error: code = Unknown desc = Error response from daemon: No such container: 2142006050fe3d6472faea03ddf0ca8a2c6684055d4a029a2fe32467d6e90142"
Sep 17 08:51:38 addons-118348 kubelet[2448]: I0917 08:51:38.495838 2448 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtpqj\" (UniqueName: \"kubernetes.io/projected/2f41b6f7-f293-467f-8215-b24af50ec8ba-kube-api-access-gtpqj\") pod \"2f41b6f7-f293-467f-8215-b24af50ec8ba\" (UID: \"2f41b6f7-f293-467f-8215-b24af50ec8ba\") "
Sep 17 08:51:38 addons-118348 kubelet[2448]: I0917 08:51:38.495898 2448 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzmhw\" (UniqueName: \"kubernetes.io/projected/29edb9a3-341b-486a-8045-5546e8911d8c-kube-api-access-bzmhw\") pod \"29edb9a3-341b-486a-8045-5546e8911d8c\" (UID: \"29edb9a3-341b-486a-8045-5546e8911d8c\") "
Sep 17 08:51:38 addons-118348 kubelet[2448]: I0917 08:51:38.497882 2448 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f41b6f7-f293-467f-8215-b24af50ec8ba-kube-api-access-gtpqj" (OuterVolumeSpecName: "kube-api-access-gtpqj") pod "2f41b6f7-f293-467f-8215-b24af50ec8ba" (UID: "2f41b6f7-f293-467f-8215-b24af50ec8ba"). InnerVolumeSpecName "kube-api-access-gtpqj". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 17 08:51:38 addons-118348 kubelet[2448]: I0917 08:51:38.497929 2448 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29edb9a3-341b-486a-8045-5546e8911d8c-kube-api-access-bzmhw" (OuterVolumeSpecName: "kube-api-access-bzmhw") pod "29edb9a3-341b-486a-8045-5546e8911d8c" (UID: "29edb9a3-341b-486a-8045-5546e8911d8c"). InnerVolumeSpecName "kube-api-access-bzmhw". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 17 08:51:38 addons-118348 kubelet[2448]: I0917 08:51:38.596273 2448 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-gtpqj\" (UniqueName: \"kubernetes.io/projected/2f41b6f7-f293-467f-8215-b24af50ec8ba-kube-api-access-gtpqj\") on node \"addons-118348\" DevicePath \"\""
Sep 17 08:51:38 addons-118348 kubelet[2448]: I0917 08:51:38.596306 2448 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-bzmhw\" (UniqueName: \"kubernetes.io/projected/29edb9a3-341b-486a-8045-5546e8911d8c-kube-api-access-bzmhw\") on node \"addons-118348\" DevicePath \"\""
==> storage-provisioner [d9dc593c31cf] <==
I0917 08:39:06.877802 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0917 08:39:06.894568 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0917 08:39:06.894620 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0917 08:39:06.986255 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0917 08:39:06.986504 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-118348_561973a3-012f-4d8f-a372-39a4ab88c714!
I0917 08:39:06.987752 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"436ede53-6305-498f-8f3a-584f57c88cbf", APIVersion:"v1", ResourceVersion:"622", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-118348_561973a3-012f-4d8f-a372-39a4ab88c714 became leader
I0917 08:39:07.092767 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-118348_561973a3-012f-4d8f-a372-39a4ab88c714!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-118348 -n addons-118348
helpers_test.go:261: (dbg) Run: kubectl --context addons-118348 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-118348 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-118348 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-118348/192.168.49.2
Start Time: Tue, 17 Sep 2024 08:42:24 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.28
IPs:
IP: 10.244.0.28
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-j84qr (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-j84qr:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m15s default-scheduler Successfully assigned default/busybox to addons-118348
Normal Pulling 7m37s (x4 over 9m14s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m37s (x4 over 9m14s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m37s (x4 over 9m14s) kubelet Error: ErrImagePull
Warning Failed 7m24s (x6 over 9m13s) kubelet Error: ImagePullBackOff
Normal BackOff 4m12s (x20 over 9m13s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (73.40s)