=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:328: registry stabilized in 2.526654ms
addons_test.go:330: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-n8x7q" [ab1423e5-b667-4a7f-96f5-061bb4596eeb] Running
addons_test.go:330: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 6.00269578s
addons_test.go:333: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-8z8jc" [db35f6da-74a1-46c1-8ca2-4c7e51bf1986] Running
addons_test.go:333: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.002843702s
addons_test.go:338: (dbg) Run: kubectl --context addons-135472 delete po -l run=registry-test --now
addons_test.go:343: (dbg) Run: kubectl --context addons-135472 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:343: (dbg) Non-zero exit: kubectl --context addons-135472 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.076617631s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:345: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-135472 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:349: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:357: (dbg) Run: out/minikube-linux-amd64 -p addons-135472 ip
2024/09/20 21:01:09 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:386: (dbg) Run: out/minikube-linux-amd64 -p addons-135472 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-135472
helpers_test.go:235: (dbg) docker inspect addons-135472:
-- stdout --
[
{
"Id": "f41c27280e30b093d965dee05d6cd1ed653e9ef1b0cd6d7ffb9de2a1c20deea3",
"Created": "2024-09-20T20:48:12.857258893Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 18381,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-20T20:48:12.975780562Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:d94335c0cd164ddebb3c5158e317bcf6d2e08dc08f448d25251f425acb842829",
"ResolvConfPath": "/var/lib/docker/containers/f41c27280e30b093d965dee05d6cd1ed653e9ef1b0cd6d7ffb9de2a1c20deea3/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/f41c27280e30b093d965dee05d6cd1ed653e9ef1b0cd6d7ffb9de2a1c20deea3/hostname",
"HostsPath": "/var/lib/docker/containers/f41c27280e30b093d965dee05d6cd1ed653e9ef1b0cd6d7ffb9de2a1c20deea3/hosts",
"LogPath": "/var/lib/docker/containers/f41c27280e30b093d965dee05d6cd1ed653e9ef1b0cd6d7ffb9de2a1c20deea3/f41c27280e30b093d965dee05d6cd1ed653e9ef1b0cd6d7ffb9de2a1c20deea3-json.log",
"Name": "/addons-135472",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-135472:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-135472",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/f209ae95c12dac22375c1b9ef598d41be2892650e39d7191bcc99fe15053ed8f-init/diff:/var/lib/docker/overlay2/467fd240a27e84496133f634bb50855964a0d9e03013662bcf99182d5b8fdb59/diff",
"MergedDir": "/var/lib/docker/overlay2/f209ae95c12dac22375c1b9ef598d41be2892650e39d7191bcc99fe15053ed8f/merged",
"UpperDir": "/var/lib/docker/overlay2/f209ae95c12dac22375c1b9ef598d41be2892650e39d7191bcc99fe15053ed8f/diff",
"WorkDir": "/var/lib/docker/overlay2/f209ae95c12dac22375c1b9ef598d41be2892650e39d7191bcc99fe15053ed8f/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "volume",
"Name": "addons-135472",
"Source": "/var/lib/docker/volumes/addons-135472/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
},
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
}
],
"Config": {
"Hostname": "addons-135472",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-135472",
"name.minikube.sigs.k8s.io": "addons-135472",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "dabe47a7a3640720971594b9815f336bc70b32d3422bfc75d6c5ffb81de3cc31",
"SandboxKey": "/var/run/docker/netns/dabe47a7a364",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-135472": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "5df1c68f89b4f25ea932b34d898870b97f752788af0a167f520f95df7e91872e",
"EndpointID": "0417389d16baf5cbfb2cad117eb13036b2da22003d59024e78bb5b9edf7bbef7",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-135472",
"f41c27280e30"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-135472 -n addons-135472
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-135472 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| start | --download-only -p | download-docker-003803 | jenkins | v1.34.0 | 20 Sep 24 20:47 UTC | |
| | download-docker-003803 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p download-docker-003803 | download-docker-003803 | jenkins | v1.34.0 | 20 Sep 24 20:47 UTC | 20 Sep 24 20:47 UTC |
| start | --download-only -p | binary-mirror-574210 | jenkins | v1.34.0 | 20 Sep 24 20:47 UTC | |
| | binary-mirror-574210 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:39611 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-574210 | binary-mirror-574210 | jenkins | v1.34.0 | 20 Sep 24 20:47 UTC | 20 Sep 24 20:47 UTC |
| addons | disable dashboard -p | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 20:47 UTC | |
| | addons-135472 | | | | | |
| addons | enable dashboard -p | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 20:47 UTC | |
| | addons-135472 | | | | | |
| start | -p addons-135472 --wait=true | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 20:47 UTC | 20 Sep 24 20:51 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| addons | addons-135472 addons disable | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 20:51 UTC | 20 Sep 24 20:51 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | addons-135472 addons disable | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | addons-135472 addons | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | -p addons-135472 | | | | | |
| addons | disable cloud-spanner -p | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | addons-135472 | | | | | |
| addons | enable headlamp | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | -p addons-135472 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-135472 ssh cat | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | /opt/local-path-provisioner/pvc-1b91b000-5e84-4be3-a317-9707e25013f8_default_test-pvc/file1 | | | | | |
| addons | addons-135472 addons disable | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | addons-135472 | | | | | |
| addons | addons-135472 addons disable | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ssh | addons-135472 ssh curl -s | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-135472 ip | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| addons | addons-135472 addons disable | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-135472 addons | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-135472 addons disable | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| addons | addons-135472 addons | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:00 UTC | 20 Sep 24 21:00 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-135472 ip | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:01 UTC | 20 Sep 24 21:01 UTC |
| addons | addons-135472 addons disable | addons-135472 | jenkins | v1.34.0 | 20 Sep 24 21:01 UTC | 20 Sep 24 21:01 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/20 20:47:49
Running on machine: ubuntu-20-agent-14
Binary: Built with gc go1.23.0 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0920 20:47:49.595195 17609 out.go:345] Setting OutFile to fd 1 ...
I0920 20:47:49.595274 17609 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0920 20:47:49.595279 17609 out.go:358] Setting ErrFile to fd 2...
I0920 20:47:49.595283 17609 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0920 20:47:49.595446 17609 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19672-9514/.minikube/bin
I0920 20:47:49.595955 17609 out.go:352] Setting JSON to false
I0920 20:47:49.596718 17609 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":1818,"bootTime":1726863452,"procs":173,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1069-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0920 20:47:49.596798 17609 start.go:139] virtualization: kvm guest
I0920 20:47:49.598829 17609 out.go:177] * [addons-135472] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0920 20:47:49.599991 17609 out.go:177] - MINIKUBE_LOCATION=19672
I0920 20:47:49.600004 17609 notify.go:220] Checking for updates...
I0920 20:47:49.602178 17609 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0920 20:47:49.603414 17609 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19672-9514/kubeconfig
I0920 20:47:49.604817 17609 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19672-9514/.minikube
I0920 20:47:49.606041 17609 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0920 20:47:49.607143 17609 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0920 20:47:49.608283 17609 driver.go:394] Setting default libvirt URI to qemu:///system
I0920 20:47:49.628337 17609 docker.go:123] docker version: linux-27.3.1:Docker Engine - Community
I0920 20:47:49.628424 17609 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0920 20:47:49.668264 17609 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-20 20:47:49.660396796 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.3.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:[WARNING: br
idge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled] ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.17.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.7] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0920 20:47:49.668394 17609 docker.go:318] overlay module found
I0920 20:47:49.670132 17609 out.go:177] * Using the docker driver based on user configuration
I0920 20:47:49.671221 17609 start.go:297] selected driver: docker
I0920 20:47:49.671233 17609 start.go:901] validating driver "docker" against <nil>
I0920 20:47:49.671247 17609 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0920 20:47:49.671967 17609 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0920 20:47:49.711685 17609 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-20 20:47:49.704083279 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.3.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:[WARNING: br
idge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled] ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.17.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.7] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0920 20:47:49.711852 17609 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0920 20:47:49.712071 17609 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0920 20:47:49.713550 17609 out.go:177] * Using Docker driver with root privileges
I0920 20:47:49.714804 17609 cni.go:84] Creating CNI manager for ""
I0920 20:47:49.714867 17609 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0920 20:47:49.714880 17609 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0920 20:47:49.714945 17609 start.go:340] cluster config:
{Name:addons-135472 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-135472 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0920 20:47:49.716124 17609 out.go:177] * Starting "addons-135472" primary control-plane node in "addons-135472" cluster
I0920 20:47:49.717136 17609 cache.go:121] Beginning downloading kic base image for docker with docker
I0920 20:47:49.718241 17609 out.go:177] * Pulling base image v0.0.45-1726784731-19672 ...
I0920 20:47:49.719194 17609 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0920 20:47:49.719221 17609 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed in local docker daemon
I0920 20:47:49.719232 17609 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19672-9514/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4
I0920 20:47:49.719239 17609 cache.go:56] Caching tarball of preloaded images
I0920 20:47:49.719319 17609 preload.go:172] Found /home/jenkins/minikube-integration/19672-9514/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0920 20:47:49.719329 17609 cache.go:59] Finished verifying existence of preloaded tar for v1.31.1 on docker
I0920 20:47:49.719615 17609 profile.go:143] Saving config to /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/config.json ...
I0920 20:47:49.719635 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/config.json: {Name:mkfde33f11021d55e33f07d8236ae59c0e285310 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:47:49.733126 17609 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed to local cache
I0920 20:47:49.733211 17609 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed in local cache directory
I0920 20:47:49.733228 17609 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed in local cache directory, skipping pull
I0920 20:47:49.733232 17609 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed exists in cache, skipping pull
I0920 20:47:49.733239 17609 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed as a tarball
I0920 20:47:49.733246 17609 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed from local cache
I0920 20:48:01.398998 17609 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed from cached tarball
I0920 20:48:01.399032 17609 cache.go:194] Successfully downloaded all kic artifacts
I0920 20:48:01.399072 17609 start.go:360] acquireMachinesLock for addons-135472: {Name:mkceb13fc1a1aab5ee2d4770834cefb10ed88226 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0920 20:48:01.399170 17609 start.go:364] duration metric: took 79.648µs to acquireMachinesLock for "addons-135472"
I0920 20:48:01.399195 17609 start.go:93] Provisioning new machine with config: &{Name:addons-135472 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-135472 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0920 20:48:01.399273 17609 start.go:125] createHost starting for "" (driver="docker")
I0920 20:48:01.401003 17609 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0920 20:48:01.401234 17609 start.go:159] libmachine.API.Create for "addons-135472" (driver="docker")
I0920 20:48:01.401274 17609 client.go:168] LocalClient.Create starting
I0920 20:48:01.401357 17609 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19672-9514/.minikube/certs/ca.pem
I0920 20:48:01.589320 17609 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19672-9514/.minikube/certs/cert.pem
I0920 20:48:01.756574 17609 cli_runner.go:164] Run: docker network inspect addons-135472 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0920 20:48:01.771856 17609 cli_runner.go:211] docker network inspect addons-135472 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0920 20:48:01.771916 17609 network_create.go:284] running [docker network inspect addons-135472] to gather additional debugging logs...
I0920 20:48:01.771932 17609 cli_runner.go:164] Run: docker network inspect addons-135472
W0920 20:48:01.785948 17609 cli_runner.go:211] docker network inspect addons-135472 returned with exit code 1
I0920 20:48:01.785970 17609 network_create.go:287] error running [docker network inspect addons-135472]: docker network inspect addons-135472: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-135472 not found
I0920 20:48:01.785983 17609 network_create.go:289] output of [docker network inspect addons-135472]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-135472 not found
** /stderr **
I0920 20:48:01.786054 17609 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0920 20:48:01.799553 17609 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001a1a9f0}
I0920 20:48:01.799590 17609 network_create.go:124] attempt to create docker network addons-135472 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0920 20:48:01.799624 17609 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-135472 addons-135472
I0920 20:48:01.854867 17609 network_create.go:108] docker network addons-135472 192.168.49.0/24 created
I0920 20:48:01.854896 17609 kic.go:121] calculated static IP "192.168.49.2" for the "addons-135472" container
I0920 20:48:01.854949 17609 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0920 20:48:01.870070 17609 cli_runner.go:164] Run: docker volume create addons-135472 --label name.minikube.sigs.k8s.io=addons-135472 --label created_by.minikube.sigs.k8s.io=true
I0920 20:48:01.885512 17609 oci.go:103] Successfully created a docker volume addons-135472
I0920 20:48:01.885585 17609 cli_runner.go:164] Run: docker run --rm --name addons-135472-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-135472 --entrypoint /usr/bin/test -v addons-135472:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed -d /var/lib
I0920 20:48:09.020783 17609 cli_runner.go:217] Completed: docker run --rm --name addons-135472-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-135472 --entrypoint /usr/bin/test -v addons-135472:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed -d /var/lib: (7.135150148s)
I0920 20:48:09.020812 17609 oci.go:107] Successfully prepared a docker volume addons-135472
I0920 20:48:09.020833 17609 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0920 20:48:09.020854 17609 kic.go:194] Starting extracting preloaded images to volume ...
I0920 20:48:09.020912 17609 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19672-9514/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-135472:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed -I lz4 -xf /preloaded.tar -C /extractDir
I0920 20:48:12.803589 17609 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19672-9514/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-135472:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed -I lz4 -xf /preloaded.tar -C /extractDir: (3.782641686s)
I0920 20:48:12.803614 17609 kic.go:203] duration metric: took 3.782758484s to extract preloaded images to volume ...
W0920 20:48:12.803728 17609 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0920 20:48:12.803880 17609 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0920 20:48:12.844031 17609 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-135472 --name addons-135472 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-135472 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-135472 --network addons-135472 --ip 192.168.49.2 --volume addons-135472:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed
I0920 20:48:13.137158 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Running}}
I0920 20:48:13.154357 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:13.171762 17609 cli_runner.go:164] Run: docker exec addons-135472 stat /var/lib/dpkg/alternatives/iptables
I0920 20:48:13.210825 17609 oci.go:144] the created container "addons-135472" has a running status.
I0920 20:48:13.210851 17609 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa...
I0920 20:48:13.297881 17609 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0920 20:48:13.315591 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:13.330304 17609 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0920 20:48:13.330323 17609 kic_runner.go:114] Args: [docker exec --privileged addons-135472 chown docker:docker /home/docker/.ssh/authorized_keys]
I0920 20:48:13.371368 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:13.385684 17609 machine.go:93] provisionDockerMachine start ...
I0920 20:48:13.385756 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:13.399880 17609 main.go:141] libmachine: Using SSH client type: native
I0920 20:48:13.400111 17609 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 20:48:13.400129 17609 main.go:141] libmachine: About to run SSH command:
hostname
I0920 20:48:13.400683 17609 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: read tcp 127.0.0.1:38192->127.0.0.1:32768: read: connection reset by peer
I0920 20:48:16.524245 17609 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-135472
I0920 20:48:16.524272 17609 ubuntu.go:169] provisioning hostname "addons-135472"
I0920 20:48:16.524328 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:16.540143 17609 main.go:141] libmachine: Using SSH client type: native
I0920 20:48:16.540296 17609 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 20:48:16.540309 17609 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-135472 && echo "addons-135472" | sudo tee /etc/hostname
I0920 20:48:16.674058 17609 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-135472
I0920 20:48:16.674118 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:16.688911 17609 main.go:141] libmachine: Using SSH client type: native
I0920 20:48:16.689071 17609 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 20:48:16.689087 17609 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-135472' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-135472/g' /etc/hosts;
else
echo '127.0.1.1 addons-135472' | sudo tee -a /etc/hosts;
fi
fi
I0920 20:48:16.812537 17609 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0920 20:48:16.812561 17609 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19672-9514/.minikube CaCertPath:/home/jenkins/minikube-integration/19672-9514/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19672-9514/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19672-9514/.minikube}
I0920 20:48:16.812592 17609 ubuntu.go:177] setting up certificates
I0920 20:48:16.812614 17609 provision.go:84] configureAuth start
I0920 20:48:16.812659 17609 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-135472
I0920 20:48:16.827061 17609 provision.go:143] copyHostCerts
I0920 20:48:16.827124 17609 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19672-9514/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19672-9514/.minikube/ca.pem (1082 bytes)
I0920 20:48:16.827234 17609 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19672-9514/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19672-9514/.minikube/cert.pem (1123 bytes)
I0920 20:48:16.827291 17609 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19672-9514/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19672-9514/.minikube/key.pem (1679 bytes)
I0920 20:48:16.827342 17609 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19672-9514/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19672-9514/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19672-9514/.minikube/certs/ca-key.pem org=jenkins.addons-135472 san=[127.0.0.1 192.168.49.2 addons-135472 localhost minikube]
I0920 20:48:17.067241 17609 provision.go:177] copyRemoteCerts
I0920 20:48:17.067290 17609 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0920 20:48:17.067322 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:17.082987 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:17.172678 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I0920 20:48:17.192362 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0920 20:48:17.211466 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I0920 20:48:17.230140 17609 provision.go:87] duration metric: took 417.510629ms to configureAuth
I0920 20:48:17.230164 17609 ubuntu.go:193] setting minikube options for container-runtime
I0920 20:48:17.230312 17609 config.go:182] Loaded profile config "addons-135472": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0920 20:48:17.230351 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:17.245397 17609 main.go:141] libmachine: Using SSH client type: native
I0920 20:48:17.245583 17609 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 20:48:17.245598 17609 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0920 20:48:17.369043 17609 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0920 20:48:17.369068 17609 ubuntu.go:71] root file system type: overlay
I0920 20:48:17.369222 17609 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0920 20:48:17.369275 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:17.384307 17609 main.go:141] libmachine: Using SSH client type: native
I0920 20:48:17.384451 17609 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 20:48:17.384506 17609 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0920 20:48:17.518528 17609 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0920 20:48:17.518608 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:17.533343 17609 main.go:141] libmachine: Using SSH client type: native
I0920 20:48:17.533536 17609 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 20:48:17.533561 17609 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0920 20:48:18.171247 17609 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-09-19 14:24:32.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-20 20:48:17.512989472 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0920 20:48:18.171272 17609 machine.go:96] duration metric: took 4.785570882s to provisionDockerMachine
I0920 20:48:18.171283 17609 client.go:171] duration metric: took 16.770001018s to LocalClient.Create
I0920 20:48:18.171298 17609 start.go:167] duration metric: took 16.770066651s to libmachine.API.Create "addons-135472"
I0920 20:48:18.171305 17609 start.go:293] postStartSetup for "addons-135472" (driver="docker")
I0920 20:48:18.171314 17609 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0920 20:48:18.171358 17609 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0920 20:48:18.171389 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:18.186559 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:18.276869 17609 ssh_runner.go:195] Run: cat /etc/os-release
I0920 20:48:18.279370 17609 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0920 20:48:18.279399 17609 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0920 20:48:18.279406 17609 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0920 20:48:18.279414 17609 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0920 20:48:18.279423 17609 filesync.go:126] Scanning /home/jenkins/minikube-integration/19672-9514/.minikube/addons for local assets ...
I0920 20:48:18.279468 17609 filesync.go:126] Scanning /home/jenkins/minikube-integration/19672-9514/.minikube/files for local assets ...
I0920 20:48:18.279491 17609 start.go:296] duration metric: took 108.180687ms for postStartSetup
I0920 20:48:18.279717 17609 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-135472
I0920 20:48:18.294318 17609 profile.go:143] Saving config to /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/config.json ...
I0920 20:48:18.294525 17609 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0920 20:48:18.294561 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:18.308963 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:18.397264 17609 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0920 20:48:18.400802 17609 start.go:128] duration metric: took 17.001517315s to createHost
I0920 20:48:18.400821 17609 start.go:83] releasing machines lock for "addons-135472", held for 17.001637211s
I0920 20:48:18.400867 17609 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-135472
I0920 20:48:18.416868 17609 ssh_runner.go:195] Run: cat /version.json
I0920 20:48:18.416906 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:18.416965 17609 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0920 20:48:18.417036 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:18.432505 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:18.433800 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:18.516201 17609 ssh_runner.go:195] Run: systemctl --version
I0920 20:48:18.584952 17609 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0920 20:48:18.588499 17609 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0920 20:48:18.608767 17609 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0920 20:48:18.608814 17609 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0920 20:48:18.631114 17609 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0920 20:48:18.631131 17609 start.go:495] detecting cgroup driver to use...
I0920 20:48:18.631159 17609 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0920 20:48:18.631257 17609 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0920 20:48:18.643481 17609 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0920 20:48:18.650963 17609 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0920 20:48:18.658231 17609 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0920 20:48:18.658278 17609 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0920 20:48:18.665986 17609 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0920 20:48:18.673571 17609 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0920 20:48:18.680962 17609 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0920 20:48:18.688614 17609 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0920 20:48:18.695951 17609 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0920 20:48:18.703681 17609 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0920 20:48:18.711443 17609 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0920 20:48:18.719266 17609 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0920 20:48:18.725887 17609 crio.go:166] couldn't verify netfilter by "sudo sysctl net.bridge.bridge-nf-call-iptables" which might be okay. error: sudo sysctl net.bridge.bridge-nf-call-iptables: Process exited with status 255
stdout:
stderr:
sysctl: cannot stat /proc/sys/net/bridge/bridge-nf-call-iptables: No such file or directory
I0920 20:48:18.725923 17609 ssh_runner.go:195] Run: sudo modprobe br_netfilter
I0920 20:48:18.737216 17609 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0920 20:48:18.743894 17609 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 20:48:18.811730 17609 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0920 20:48:18.889082 17609 start.go:495] detecting cgroup driver to use...
I0920 20:48:18.889128 17609 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0920 20:48:18.889192 17609 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0920 20:48:18.898926 17609 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0920 20:48:18.898983 17609 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0920 20:48:18.908643 17609 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0920 20:48:18.922384 17609 ssh_runner.go:195] Run: which cri-dockerd
I0920 20:48:18.925128 17609 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0920 20:48:18.932915 17609 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0920 20:48:18.948226 17609 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0920 20:48:19.029786 17609 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0920 20:48:19.111221 17609 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0920 20:48:19.111352 17609 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0920 20:48:19.125682 17609 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 20:48:19.193033 17609 ssh_runner.go:195] Run: sudo systemctl restart docker
I0920 20:48:19.421870 17609 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0920 20:48:19.431781 17609 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0920 20:48:19.441015 17609 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0920 20:48:19.515978 17609 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0920 20:48:19.587593 17609 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 20:48:19.659324 17609 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0920 20:48:19.669757 17609 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0920 20:48:19.678122 17609 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 20:48:19.747744 17609 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0920 20:48:19.801221 17609 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0920 20:48:19.801294 17609 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0920 20:48:19.804335 17609 start.go:563] Will wait 60s for crictl version
I0920 20:48:19.804382 17609 ssh_runner.go:195] Run: which crictl
I0920 20:48:19.807201 17609 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0920 20:48:19.835847 17609 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.3.0
RuntimeApiVersion: v1
I0920 20:48:19.835890 17609 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0920 20:48:19.856171 17609 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0920 20:48:19.876877 17609 out.go:235] * Preparing Kubernetes v1.31.1 on Docker 27.3.0 ...
I0920 20:48:19.876957 17609 cli_runner.go:164] Run: docker network inspect addons-135472 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0920 20:48:19.891947 17609 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0920 20:48:19.894906 17609 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0920 20:48:19.904191 17609 kubeadm.go:883] updating cluster {Name:addons-135472 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-135472 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0920 20:48:19.904300 17609 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0920 20:48:19.904345 17609 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0920 20:48:19.920505 17609 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0920 20:48:19.920522 17609 docker.go:615] Images already preloaded, skipping extraction
I0920 20:48:19.920569 17609 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0920 20:48:19.935725 17609 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0920 20:48:19.935747 17609 cache_images.go:84] Images are preloaded, skipping loading
I0920 20:48:19.935759 17609 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.1 docker true true} ...
I0920 20:48:19.935869 17609 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-135472 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.1 ClusterName:addons-135472 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0920 20:48:19.935918 17609 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0920 20:48:19.975115 17609 cni.go:84] Creating CNI manager for ""
I0920 20:48:19.975139 17609 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0920 20:48:19.975148 17609 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0920 20:48:19.975168 17609 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-135472 NodeName:addons-135472 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0920 20:48:19.975326 17609 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-135472"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0920 20:48:19.975388 17609 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.1
I0920 20:48:19.982751 17609 binaries.go:44] Found k8s binaries, skipping transfer
I0920 20:48:19.982812 17609 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0920 20:48:19.989828 17609 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0920 20:48:20.004244 17609 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0920 20:48:20.018176 17609 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0920 20:48:20.032748 17609 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0920 20:48:20.035627 17609 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0920 20:48:20.044378 17609 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 20:48:20.112623 17609 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0920 20:48:20.123229 17609 certs.go:68] Setting up /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472 for IP: 192.168.49.2
I0920 20:48:20.123243 17609 certs.go:194] generating shared ca certs ...
I0920 20:48:20.123257 17609 certs.go:226] acquiring lock for ca certs: {Name:mke171823f01199a0b3b7794b5263fc14bd774ba Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.123357 17609 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19672-9514/.minikube/ca.key
I0920 20:48:20.375201 17609 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19672-9514/.minikube/ca.crt ...
I0920 20:48:20.375224 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/ca.crt: {Name:mkde573a47ae4b9856c76951253e219432b4eacd Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.375361 17609 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19672-9514/.minikube/ca.key ...
I0920 20:48:20.375370 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/ca.key: {Name:mk5da5948f111a983464b660fda183591ee045d5 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.375438 17609 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19672-9514/.minikube/proxy-client-ca.key
I0920 20:48:20.416140 17609 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19672-9514/.minikube/proxy-client-ca.crt ...
I0920 20:48:20.416158 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/proxy-client-ca.crt: {Name:mk66f7340eaef942dd46c29e9b11fc3e8e281e80 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.416260 17609 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19672-9514/.minikube/proxy-client-ca.key ...
I0920 20:48:20.416269 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/proxy-client-ca.key: {Name:mk8b151339bb4f28791e62f764eef3ab92c0dc48 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.416328 17609 certs.go:256] generating profile certs ...
I0920 20:48:20.416373 17609 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/client.key
I0920 20:48:20.416387 17609 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/client.crt with IP's: []
I0920 20:48:20.550916 17609 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/client.crt ...
I0920 20:48:20.550939 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/client.crt: {Name:mk3a2f3373d9f9f36ab1e1e7e369e2a555d1d88e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.551062 17609 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/client.key ...
I0920 20:48:20.551073 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/client.key: {Name:mkfdc35a57265d3d42d627cc325844d768ba3ab6 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.551136 17609 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.key.247b41b2
I0920 20:48:20.551153 17609 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.crt.247b41b2 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0920 20:48:20.714474 17609 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.crt.247b41b2 ...
I0920 20:48:20.714491 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.crt.247b41b2: {Name:mk5c0a743de50447e0822476f4f8dc2a5d7a95a9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.714598 17609 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.key.247b41b2 ...
I0920 20:48:20.714609 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.key.247b41b2: {Name:mkfc5dd59d8f4f92c08b559ca341314acca771c1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.714671 17609 certs.go:381] copying /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.crt.247b41b2 -> /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.crt
I0920 20:48:20.714736 17609 certs.go:385] copying /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.key.247b41b2 -> /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.key
I0920 20:48:20.714780 17609 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/proxy-client.key
I0920 20:48:20.714795 17609 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/proxy-client.crt with IP's: []
I0920 20:48:20.822924 17609 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/proxy-client.crt ...
I0920 20:48:20.822943 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/proxy-client.crt: {Name:mka900140cba3b2c5393e2ede0869264da1a04e6 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.823061 17609 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/proxy-client.key ...
I0920 20:48:20.823073 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/proxy-client.key: {Name:mkd2f80bf4c8cdca636f9a2dfe3e391857631ed1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:20.823276 17609 certs.go:484] found cert: /home/jenkins/minikube-integration/19672-9514/.minikube/certs/ca-key.pem (1679 bytes)
I0920 20:48:20.823305 17609 certs.go:484] found cert: /home/jenkins/minikube-integration/19672-9514/.minikube/certs/ca.pem (1082 bytes)
I0920 20:48:20.823340 17609 certs.go:484] found cert: /home/jenkins/minikube-integration/19672-9514/.minikube/certs/cert.pem (1123 bytes)
I0920 20:48:20.823375 17609 certs.go:484] found cert: /home/jenkins/minikube-integration/19672-9514/.minikube/certs/key.pem (1679 bytes)
I0920 20:48:20.824108 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0920 20:48:20.844285 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0920 20:48:20.862887 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0920 20:48:20.881346 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0920 20:48:20.900059 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0920 20:48:20.919341 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0920 20:48:20.938260 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0920 20:48:20.957297 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/profiles/addons-135472/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0920 20:48:20.976071 17609 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19672-9514/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0920 20:48:20.995511 17609 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0920 20:48:21.010955 17609 ssh_runner.go:195] Run: openssl version
I0920 20:48:21.015418 17609 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0920 20:48:21.023147 17609 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0920 20:48:21.026020 17609 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 20 20:48 /usr/share/ca-certificates/minikubeCA.pem
I0920 20:48:21.026062 17609 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0920 20:48:21.031637 17609 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0920 20:48:21.038860 17609 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0920 20:48:21.041418 17609 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0920 20:48:21.041456 17609 kubeadm.go:392] StartCluster: {Name:addons-135472 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726784731-19672@sha256:7f8c62ddb0100a5b958dd19c5b5478b8c7ef13da9a0a4d6c7d18f43544e0dbed Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-135472 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0920 20:48:21.041571 17609 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0920 20:48:21.056504 17609 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0920 20:48:21.063238 17609 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0920 20:48:21.070170 17609 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0920 20:48:21.070206 17609 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0920 20:48:21.076912 17609 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0920 20:48:21.076925 17609 kubeadm.go:157] found existing configuration files:
I0920 20:48:21.076952 17609 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0920 20:48:21.083514 17609 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0920 20:48:21.083552 17609 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0920 20:48:21.090077 17609 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0920 20:48:21.096630 17609 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0920 20:48:21.096668 17609 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0920 20:48:21.102972 17609 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0920 20:48:21.109577 17609 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0920 20:48:21.109614 17609 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0920 20:48:21.116014 17609 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0920 20:48:21.122802 17609 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0920 20:48:21.122832 17609 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0920 20:48:21.130046 17609 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0920 20:48:21.161311 17609 kubeadm.go:310] [init] Using Kubernetes version: v1.31.1
I0920 20:48:21.161374 17609 kubeadm.go:310] [preflight] Running pre-flight checks
I0920 20:48:21.178871 17609 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0920 20:48:21.178950 17609 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1069-gcp[0m
I0920 20:48:21.178999 17609 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0920 20:48:21.179073 17609 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0920 20:48:21.179147 17609 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0920 20:48:21.179231 17609 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0920 20:48:21.179303 17609 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0920 20:48:21.179376 17609 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0920 20:48:21.179418 17609 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0920 20:48:21.179457 17609 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0920 20:48:21.179499 17609 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0920 20:48:21.179548 17609 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0920 20:48:21.223726 17609 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0920 20:48:21.223865 17609 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0920 20:48:21.224002 17609 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0920 20:48:21.232797 17609 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0920 20:48:21.235307 17609 out.go:235] - Generating certificates and keys ...
I0920 20:48:21.235406 17609 kubeadm.go:310] [certs] Using existing ca certificate authority
I0920 20:48:21.235510 17609 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0920 20:48:21.341616 17609 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0920 20:48:21.473249 17609 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0920 20:48:21.572138 17609 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0920 20:48:21.888300 17609 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0920 20:48:21.986111 17609 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0920 20:48:21.986242 17609 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-135472 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0920 20:48:22.150529 17609 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0920 20:48:22.150680 17609 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-135472 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0920 20:48:22.582891 17609 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0920 20:48:22.934736 17609 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0920 20:48:23.050268 17609 kubeadm.go:310] [certs] Generating "sa" key and public key
I0920 20:48:23.050369 17609 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0920 20:48:23.263098 17609 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0920 20:48:23.399518 17609 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0920 20:48:23.551643 17609 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0920 20:48:23.614385 17609 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0920 20:48:23.903381 17609 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0920 20:48:23.903865 17609 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0920 20:48:23.906169 17609 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0920 20:48:23.908369 17609 out.go:235] - Booting up control plane ...
I0920 20:48:23.908492 17609 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0920 20:48:23.908576 17609 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0920 20:48:23.908651 17609 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0920 20:48:23.919872 17609 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0920 20:48:23.925188 17609 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0920 20:48:23.925276 17609 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0920 20:48:24.002095 17609 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0920 20:48:24.002251 17609 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0920 20:48:25.003839 17609 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.001682763s
I0920 20:48:25.003943 17609 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0920 20:48:29.004985 17609 kubeadm.go:310] [api-check] The API server is healthy after 4.001248073s
I0920 20:48:29.026516 17609 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0920 20:48:29.035098 17609 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0920 20:48:29.048606 17609 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0920 20:48:29.048884 17609 kubeadm.go:310] [mark-control-plane] Marking the node addons-135472 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0920 20:48:29.054777 17609 kubeadm.go:310] [bootstrap-token] Using token: yog870.4e36o6rp1fgnt41w
I0920 20:48:29.056275 17609 out.go:235] - Configuring RBAC rules ...
I0920 20:48:29.056417 17609 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0920 20:48:29.058672 17609 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0920 20:48:29.063316 17609 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0920 20:48:29.065318 17609 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0920 20:48:29.067219 17609 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0920 20:48:29.069924 17609 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0920 20:48:29.410081 17609 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0920 20:48:29.826041 17609 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0920 20:48:30.408968 17609 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0920 20:48:30.409659 17609 kubeadm.go:310]
I0920 20:48:30.409776 17609 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0920 20:48:30.409795 17609 kubeadm.go:310]
I0920 20:48:30.409885 17609 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0920 20:48:30.409896 17609 kubeadm.go:310]
I0920 20:48:30.409917 17609 kubeadm.go:310] mkdir -p $HOME/.kube
I0920 20:48:30.409975 17609 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0920 20:48:30.410022 17609 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0920 20:48:30.410028 17609 kubeadm.go:310]
I0920 20:48:30.410087 17609 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0920 20:48:30.410094 17609 kubeadm.go:310]
I0920 20:48:30.410133 17609 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0920 20:48:30.410150 17609 kubeadm.go:310]
I0920 20:48:30.410235 17609 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0920 20:48:30.410343 17609 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0920 20:48:30.410430 17609 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0920 20:48:30.410441 17609 kubeadm.go:310]
I0920 20:48:30.410561 17609 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0920 20:48:30.410663 17609 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0920 20:48:30.410672 17609 kubeadm.go:310]
I0920 20:48:30.410796 17609 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token yog870.4e36o6rp1fgnt41w \
I0920 20:48:30.410881 17609 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:f491762512938f72e6c8d2a7fb1aa6e0bfc1ffa5d0eb3a7bd12200f4fb3d9bd7 \
I0920 20:48:30.410899 17609 kubeadm.go:310] --control-plane
I0920 20:48:30.410906 17609 kubeadm.go:310]
I0920 20:48:30.411010 17609 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0920 20:48:30.411021 17609 kubeadm.go:310]
I0920 20:48:30.411126 17609 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token yog870.4e36o6rp1fgnt41w \
I0920 20:48:30.411251 17609 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:f491762512938f72e6c8d2a7fb1aa6e0bfc1ffa5d0eb3a7bd12200f4fb3d9bd7
I0920 20:48:30.412555 17609 kubeadm.go:310] W0920 20:48:21.159026 1926 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0920 20:48:30.412816 17609 kubeadm.go:310] W0920 20:48:21.159550 1926 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0920 20:48:30.412990 17609 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1069-gcp\n", err: exit status 1
I0920 20:48:30.413113 17609 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0920 20:48:30.413139 17609 cni.go:84] Creating CNI manager for ""
I0920 20:48:30.413159 17609 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0920 20:48:30.415033 17609 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0920 20:48:30.417760 17609 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0920 20:48:30.425472 17609 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0920 20:48:30.440471 17609 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0920 20:48:30.440529 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 20:48:30.440555 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-135472 minikube.k8s.io/updated_at=2024_09_20T20_48_30_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=b921bee7dddd4990dd76a4773b23d7ec11e6144f minikube.k8s.io/name=addons-135472 minikube.k8s.io/primary=true
I0920 20:48:30.446893 17609 ops.go:34] apiserver oom_adj: -16
I0920 20:48:30.506221 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 20:48:31.006863 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 20:48:31.506401 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 20:48:32.006877 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 20:48:32.506272 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 20:48:33.006765 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 20:48:33.506709 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 20:48:34.007005 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 20:48:34.506475 17609 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 20:48:34.564756 17609 kubeadm.go:1113] duration metric: took 4.124274043s to wait for elevateKubeSystemPrivileges
I0920 20:48:34.564796 17609 kubeadm.go:394] duration metric: took 13.523341077s to StartCluster
I0920 20:48:34.564817 17609 settings.go:142] acquiring lock: {Name:mk599dae52b8e72abfd50bf7fe2ec2d4b59104d1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:34.564911 17609 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19672-9514/kubeconfig
I0920 20:48:34.565246 17609 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19672-9514/kubeconfig: {Name:mk83fe77f0521522a623481e5a97162528173507 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 20:48:34.565410 17609 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0920 20:48:34.565414 17609 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0920 20:48:34.565485 17609 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0920 20:48:34.565614 17609 addons.go:69] Setting yakd=true in profile "addons-135472"
I0920 20:48:34.565617 17609 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-135472"
I0920 20:48:34.565627 17609 config.go:182] Loaded profile config "addons-135472": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0920 20:48:34.565650 17609 addons.go:234] Setting addon yakd=true in "addons-135472"
I0920 20:48:34.565669 17609 addons.go:69] Setting cloud-spanner=true in profile "addons-135472"
I0920 20:48:34.565676 17609 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-135472"
I0920 20:48:34.565680 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.565686 17609 addons.go:234] Setting addon cloud-spanner=true in "addons-135472"
I0920 20:48:34.565700 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.565710 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.565786 17609 addons.go:69] Setting metrics-server=true in profile "addons-135472"
I0920 20:48:34.565823 17609 addons.go:234] Setting addon metrics-server=true in "addons-135472"
I0920 20:48:34.565853 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.565926 17609 addons.go:69] Setting storage-provisioner=true in profile "addons-135472"
I0920 20:48:34.565952 17609 addons.go:234] Setting addon storage-provisioner=true in "addons-135472"
I0920 20:48:34.565972 17609 addons.go:69] Setting volcano=true in profile "addons-135472"
I0920 20:48:34.566014 17609 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-135472"
I0920 20:48:34.566033 17609 addons.go:234] Setting addon volcano=true in "addons-135472"
I0920 20:48:34.566042 17609 addons.go:69] Setting registry=true in profile "addons-135472"
I0920 20:48:34.566057 17609 addons.go:234] Setting addon registry=true in "addons-135472"
I0920 20:48:34.566078 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.566078 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.566251 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.566320 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.566344 17609 addons.go:69] Setting volumesnapshots=true in profile "addons-135472"
I0920 20:48:34.566368 17609 addons.go:234] Setting addon volumesnapshots=true in "addons-135472"
I0920 20:48:34.566394 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.566470 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.566534 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.566792 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.566928 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.566321 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.566035 17609 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-135472"
I0920 20:48:34.567318 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.567369 17609 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-135472"
I0920 20:48:34.567397 17609 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-135472"
I0920 20:48:34.567696 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.567837 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.566004 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.568301 17609 out.go:177] * Verifying Kubernetes components...
I0920 20:48:34.568418 17609 addons.go:69] Setting ingress-dns=true in profile "addons-135472"
I0920 20:48:34.568640 17609 addons.go:234] Setting addon ingress-dns=true in "addons-135472"
I0920 20:48:34.568676 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.569291 17609 addons.go:69] Setting inspektor-gadget=true in profile "addons-135472"
I0920 20:48:34.569318 17609 addons.go:234] Setting addon inspektor-gadget=true in "addons-135472"
I0920 20:48:34.569351 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.569603 17609 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 20:48:34.569811 17609 addons.go:69] Setting gcp-auth=true in profile "addons-135472"
I0920 20:48:34.569835 17609 mustload.go:65] Loading cluster: addons-135472
I0920 20:48:34.569931 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.570008 17609 config.go:182] Loaded profile config "addons-135472": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0920 20:48:34.570254 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.571772 17609 addons.go:69] Setting default-storageclass=true in profile "addons-135472"
I0920 20:48:34.571820 17609 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-135472"
I0920 20:48:34.571905 17609 addons.go:69] Setting ingress=true in profile "addons-135472"
I0920 20:48:34.571939 17609 addons.go:234] Setting addon ingress=true in "addons-135472"
I0920 20:48:34.571986 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.569235 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.593960 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.594401 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.594978 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.602430 17609 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0920 20:48:34.605983 17609 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0920 20:48:34.606006 17609 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0920 20:48:34.606060 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.610453 17609 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0920 20:48:34.611656 17609 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0920 20:48:34.612772 17609 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0920 20:48:34.613882 17609 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0920 20:48:34.614977 17609 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0920 20:48:34.616146 17609 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0920 20:48:34.617189 17609 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0920 20:48:34.618339 17609 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0920 20:48:34.619336 17609 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0920 20:48:34.619359 17609 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0920 20:48:34.619410 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.621014 17609 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0920 20:48:34.621999 17609 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0920 20:48:34.622017 17609 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0920 20:48:34.622063 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.623662 17609 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0920 20:48:34.624954 17609 out.go:177] - Using image docker.io/registry:2.8.3
I0920 20:48:34.626021 17609 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0920 20:48:34.627043 17609 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0920 20:48:34.628125 17609 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0920 20:48:34.628141 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0920 20:48:34.628190 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.629342 17609 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0920 20:48:34.630002 17609 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0920 20:48:34.631392 17609 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0920 20:48:34.631408 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0920 20:48:34.631452 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.643376 17609 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0920 20:48:34.643406 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0920 20:48:34.643577 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.648489 17609 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0920 20:48:34.649669 17609 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0920 20:48:34.649745 17609 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0920 20:48:34.649762 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0920 20:48:34.649827 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.651120 17609 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0920 20:48:34.651141 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0920 20:48:34.651193 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.652602 17609 addons.go:234] Setting addon default-storageclass=true in "addons-135472"
I0920 20:48:34.652609 17609 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-135472"
I0920 20:48:34.652630 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.652654 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.653006 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.653236 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:34.656799 17609 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0920 20:48:34.658966 17609 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0920 20:48:34.658984 17609 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0920 20:48:34.659030 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.661631 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:34.664718 17609 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0920 20:48:34.665807 17609 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0920 20:48:34.665864 17609 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0920 20:48:34.667281 17609 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0920 20:48:34.667313 17609 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0920 20:48:34.667324 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0920 20:48:34.667370 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.668562 17609 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0920 20:48:34.668578 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0920 20:48:34.668655 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.674350 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.682432 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.686639 17609 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0920 20:48:34.687671 17609 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0920 20:48:34.687689 17609 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0920 20:48:34.687738 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.705056 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.715840 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.718518 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.721207 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.723049 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.724655 17609 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0920 20:48:34.724670 17609 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0920 20:48:34.724721 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.730039 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.731409 17609 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0920 20:48:34.732645 17609 out.go:177] - Using image docker.io/busybox:stable
I0920 20:48:34.732656 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.733858 17609 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0920 20:48:34.733880 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0920 20:48:34.733925 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:34.735101 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.737313 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.738674 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.743716 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
W0920 20:48:34.748390 17609 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0920 20:48:34.748416 17609 retry.go:31] will retry after 138.91814ms: ssh: handshake failed: EOF
W0920 20:48:34.748496 17609 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0920 20:48:34.748507 17609 retry.go:31] will retry after 344.583853ms: ssh: handshake failed: EOF
I0920 20:48:34.750510 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:34.769075 17609 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0920 20:48:34.769181 17609 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0920 20:48:34.958394 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0920 20:48:35.059448 17609 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0920 20:48:35.059519 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0920 20:48:35.066023 17609 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0920 20:48:35.066102 17609 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0920 20:48:35.069929 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0920 20:48:35.072819 17609 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0920 20:48:35.072878 17609 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0920 20:48:35.178052 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0920 20:48:35.260057 17609 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0920 20:48:35.260140 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0920 20:48:35.263246 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0920 20:48:35.263558 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0920 20:48:35.268567 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0920 20:48:35.273752 17609 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0920 20:48:35.273806 17609 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0920 20:48:35.275175 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0920 20:48:35.278733 17609 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0920 20:48:35.278782 17609 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0920 20:48:35.460109 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0920 20:48:35.560825 17609 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0920 20:48:35.561072 17609 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0920 20:48:35.561042 17609 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0920 20:48:35.561213 17609 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0920 20:48:35.563768 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0920 20:48:35.566244 17609 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0920 20:48:35.566263 17609 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0920 20:48:35.759254 17609 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0920 20:48:35.759349 17609 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0920 20:48:35.878074 17609 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0920 20:48:35.878152 17609 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0920 20:48:35.963522 17609 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0920 20:48:35.963609 17609 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0920 20:48:36.160247 17609 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0920 20:48:36.160334 17609 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0920 20:48:36.179661 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0920 20:48:36.262143 17609 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.492937609s)
I0920 20:48:36.263284 17609 node_ready.go:35] waiting up to 6m0s for node "addons-135472" to be "Ready" ...
I0920 20:48:36.263473 17609 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.494369796s)
I0920 20:48:36.263615 17609 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0920 20:48:36.269896 17609 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0920 20:48:36.269980 17609 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0920 20:48:36.271742 17609 node_ready.go:49] node "addons-135472" has status "Ready":"True"
I0920 20:48:36.271801 17609 node_ready.go:38] duration metric: took 8.43969ms for node "addons-135472" to be "Ready" ...
I0920 20:48:36.271823 17609 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0920 20:48:36.280350 17609 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-56tpz" in "kube-system" namespace to be "Ready" ...
I0920 20:48:36.468027 17609 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0920 20:48:36.468108 17609 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0920 20:48:36.767906 17609 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-135472" context rescaled to 1 replicas
I0920 20:48:36.774489 17609 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0920 20:48:36.774571 17609 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0920 20:48:36.858004 17609 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0920 20:48:36.858077 17609 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0920 20:48:36.871324 17609 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0920 20:48:36.871353 17609 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0920 20:48:36.974931 17609 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0920 20:48:36.974962 17609 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0920 20:48:37.167483 17609 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0920 20:48:37.167563 17609 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0920 20:48:37.278092 17609 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0920 20:48:37.278166 17609 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0920 20:48:37.363983 17609 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0920 20:48:37.364061 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0920 20:48:37.475827 17609 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0920 20:48:37.475926 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0920 20:48:37.659209 17609 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0920 20:48:37.659294 17609 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0920 20:48:37.766169 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0920 20:48:37.770831 17609 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0920 20:48:37.770858 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0920 20:48:37.874644 17609 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0920 20:48:37.874671 17609 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0920 20:48:38.073388 17609 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0920 20:48:38.073418 17609 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0920 20:48:38.259194 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0920 20:48:38.368016 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-56tpz" in "kube-system" namespace has status "Ready":"False"
I0920 20:48:38.460475 17609 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0920 20:48:38.460506 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0920 20:48:38.762492 17609 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0920 20:48:38.762524 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0920 20:48:38.881924 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0920 20:48:38.969423 17609 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0920 20:48:38.969454 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0920 20:48:39.759878 17609 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0920 20:48:39.759935 17609 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0920 20:48:40.369315 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0920 20:48:40.864134 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-56tpz" in "kube-system" namespace has status "Ready":"False"
I0920 20:48:41.672067 17609 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0920 20:48:41.672135 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:41.692146 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:42.371297 17609 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0920 20:48:42.661493 17609 addons.go:234] Setting addon gcp-auth=true in "addons-135472"
I0920 20:48:42.661590 17609 host.go:66] Checking if "addons-135472" exists ...
I0920 20:48:42.662157 17609 cli_runner.go:164] Run: docker container inspect addons-135472 --format={{.State.Status}}
I0920 20:48:42.689456 17609 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0920 20:48:42.689506 17609 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-135472
I0920 20:48:42.704663 17609 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19672-9514/.minikube/machines/addons-135472/id_rsa Username:docker}
I0920 20:48:42.866432 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-56tpz" in "kube-system" namespace has status "Ready":"False"
I0920 20:48:44.871064 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-56tpz" in "kube-system" namespace has status "Ready":"False"
I0920 20:48:45.680251 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (10.610283579s)
I0920 20:48:45.680265 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (10.72182263s)
I0920 20:48:45.680327 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (10.502174549s)
I0920 20:48:45.680380 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (10.416765041s)
I0920 20:48:45.680468 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (10.417198836s)
I0920 20:48:45.680620 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (10.412030752s)
I0920 20:48:45.680704 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (10.405480513s)
I0920 20:48:45.680794 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (10.220597582s)
I0920 20:48:45.680966 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (10.117171814s)
I0920 20:48:45.681015 17609 addons.go:475] Verifying addon registry=true in "addons-135472"
I0920 20:48:45.681074 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (9.501383218s)
I0920 20:48:45.681098 17609 addons.go:475] Verifying addon metrics-server=true in "addons-135472"
I0920 20:48:45.681249 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (7.422019813s)
W0920 20:48:45.681281 17609 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0920 20:48:45.681303 17609 addons.go:475] Verifying addon ingress=true in "addons-135472"
I0920 20:48:45.681305 17609 retry.go:31] will retry after 127.870382ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0920 20:48:45.681151 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (7.914895514s)
I0920 20:48:45.681371 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (6.799419166s)
I0920 20:48:45.682626 17609 out.go:177] * Verifying registry addon...
I0920 20:48:45.758698 17609 out.go:177] * Verifying ingress addon...
I0920 20:48:45.758749 17609 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-135472 service yakd-dashboard -n yakd-dashboard
I0920 20:48:45.760943 17609 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0920 20:48:45.762766 17609 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0920 20:48:45.765790 17609 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0920 20:48:45.766393 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
W0920 20:48:45.766154 17609 out.go:270] ! Enabling 'storage-provisioner-rancher' returned an error: running callbacks: [Error making local-path the default storage class: Error while marking storage class local-path as default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
I0920 20:48:45.809996 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0920 20:48:45.875963 17609 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0920 20:48:45.876040 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:46.265587 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:46.266869 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:46.765926 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:46.768108 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:47.264264 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:47.267593 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:47.285707 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-56tpz" in "kube-system" namespace has status "Ready":"False"
I0920 20:48:47.567446 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (7.198029717s)
I0920 20:48:47.567678 17609 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-135472"
I0920 20:48:47.567638 17609 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (4.878156713s)
I0920 20:48:47.569064 17609 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0920 20:48:47.569185 17609 out.go:177] * Verifying csi-hostpath-driver addon...
I0920 20:48:47.570317 17609 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0920 20:48:47.571328 17609 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0920 20:48:47.571345 17609 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0920 20:48:47.571368 17609 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0920 20:48:47.578199 17609 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0920 20:48:47.578231 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:47.593800 17609 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0920 20:48:47.593819 17609 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0920 20:48:47.669255 17609 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0920 20:48:47.669280 17609 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0920 20:48:47.689314 17609 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0920 20:48:47.764469 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:47.766871 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:48.075579 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:48.264637 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:48.266720 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:48.286936 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.476896466s)
I0920 20:48:48.576270 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:48.764590 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:48.766773 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:48.896015 17609 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.206658245s)
I0920 20:48:48.898081 17609 addons.go:475] Verifying addon gcp-auth=true in "addons-135472"
I0920 20:48:48.899472 17609 out.go:177] * Verifying gcp-auth addon...
I0920 20:48:48.901430 17609 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0920 20:48:48.958594 17609 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0920 20:48:49.076069 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:49.264215 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:49.266158 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:49.287629 17609 pod_ready.go:98] pod "coredns-7c65d6cfc9-56tpz" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 20:48:49 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 20:48:35 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 20:48:35 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 20:48:35 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 20:48:35 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.49.2 HostIPs:[{IP:192.168.49.2
}] PodIP:10.244.0.2 PodIPs:[{IP:10.244.0.2}] StartTime:2024-09-20 20:48:35 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-09-20 20:48:38 +0000 UTC,FinishedAt:2024-09-20 20:48:48 +0000 UTC,ContainerID:docker://c27c477250702f87f29cc8f546ca3d4525085e182a8ae81d81ac9e05d6fcec4b,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.3 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e ContainerID:docker://c27c477250702f87f29cc8f546ca3d4525085e182a8ae81d81ac9e05d6fcec4b Started:0xc0014161d0 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0xc000883430} {Name:kube-api-access-h9w89 MountPath:/var/run/secrets/kubernetes.io/serviceaccount
ReadOnly:true RecursiveReadOnly:0xc000883440}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
I0920 20:48:49.287656 17609 pod_ready.go:82] duration metric: took 13.007248912s for pod "coredns-7c65d6cfc9-56tpz" in "kube-system" namespace to be "Ready" ...
E0920 20:48:49.287670 17609 pod_ready.go:67] WaitExtra: waitPodCondition: pod "coredns-7c65d6cfc9-56tpz" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 20:48:49 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 20:48:35 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 20:48:35 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 20:48:35 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 20:48:35 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.4
9.2 HostIPs:[{IP:192.168.49.2}] PodIP:10.244.0.2 PodIPs:[{IP:10.244.0.2}] StartTime:2024-09-20 20:48:35 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-09-20 20:48:38 +0000 UTC,FinishedAt:2024-09-20 20:48:48 +0000 UTC,ContainerID:docker://c27c477250702f87f29cc8f546ca3d4525085e182a8ae81d81ac9e05d6fcec4b,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.3 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e ContainerID:docker://c27c477250702f87f29cc8f546ca3d4525085e182a8ae81d81ac9e05d6fcec4b Started:0xc0014161d0 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0xc000883430} {Name:kube-api-access-h9w89 MountPath:/var/run/secrets
/kubernetes.io/serviceaccount ReadOnly:true RecursiveReadOnly:0xc000883440}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
I0920 20:48:49.287682 17609 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace to be "Ready" ...
I0920 20:48:49.575593 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:49.765016 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:49.766033 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:50.075905 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:50.264275 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:50.266361 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:50.575600 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:50.765637 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:50.767017 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:51.075885 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:51.266629 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:51.267542 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:51.291812 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:48:51.576422 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:51.764306 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:51.766564 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:52.075315 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:52.264300 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:52.266178 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:52.575722 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:52.789716 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:52.790137 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:53.075757 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:53.264165 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:53.266347 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:53.293119 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:48:53.575763 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:53.763937 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:53.765813 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:54.075671 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:54.264842 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:54.266617 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:54.575377 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:54.765041 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:54.766180 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:55.075908 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:55.264729 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:55.266441 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:55.574807 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:55.764704 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:55.765976 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:55.792794 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:48:56.076201 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:56.264488 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:56.266758 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:56.575131 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:56.764414 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:56.766325 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:57.076114 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:57.264223 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:57.266466 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:57.575517 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:57.764381 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:57.766174 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:58.074678 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:58.265308 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:58.265841 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:58.291913 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:48:58.575385 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:58.765256 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:58.770408 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:59.076416 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:59.265105 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:59.267139 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:48:59.576029 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:48:59.764372 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:48:59.766664 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:00.075474 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:00.264648 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:00.266437 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:00.576024 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:00.765038 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:00.766211 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:00.793326 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:49:01.075890 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:01.263923 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:01.266054 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:01.575303 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:01.764064 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:01.765914 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:02.075457 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:02.265110 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:02.265685 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:02.575456 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:02.765031 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:02.766111 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:03.075880 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:03.263781 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:03.266275 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:03.292949 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:49:03.577751 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:03.764827 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:03.765961 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:04.076632 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:04.264729 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:04.266831 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:04.585079 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:04.764489 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:04.766157 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:05.075360 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:05.264668 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:05.265662 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:05.575359 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:05.764151 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 20:49:05.766409 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:05.792805 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:49:06.075734 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:06.264466 17609 kapi.go:107] duration metric: took 20.503519899s to wait for kubernetes.io/minikube-addons=registry ...
I0920 20:49:06.267308 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:06.574532 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:06.766856 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:07.076163 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:07.266621 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:07.575073 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:07.766823 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:08.075658 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:08.266666 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:08.292260 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:49:08.575663 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:08.767165 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:09.076537 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:09.266786 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:09.575169 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:09.767122 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:10.076156 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:10.266449 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:10.313307 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:49:10.633903 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:10.766440 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:11.075298 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:11.270862 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:11.575651 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:11.767047 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:12.076031 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:12.266622 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:12.681313 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:12.766335 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:12.792527 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:49:13.076521 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:13.267001 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:13.575241 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:13.766752 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:14.075409 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:14.267078 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:14.575780 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:14.766105 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:15.076026 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:15.301901 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:15.364418 17609 pod_ready.go:103] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"False"
I0920 20:49:15.575817 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:15.767762 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:16.075960 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:16.267233 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:16.577969 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:16.767221 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:16.792767 17609 pod_ready.go:93] pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace has status "Ready":"True"
I0920 20:49:16.792793 17609 pod_ready.go:82] duration metric: took 27.505098882s for pod "coredns-7c65d6cfc9-7m4lj" in "kube-system" namespace to be "Ready" ...
I0920 20:49:16.792807 17609 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-135472" in "kube-system" namespace to be "Ready" ...
I0920 20:49:16.796962 17609 pod_ready.go:93] pod "etcd-addons-135472" in "kube-system" namespace has status "Ready":"True"
I0920 20:49:16.796979 17609 pod_ready.go:82] duration metric: took 4.16436ms for pod "etcd-addons-135472" in "kube-system" namespace to be "Ready" ...
I0920 20:49:16.796988 17609 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-135472" in "kube-system" namespace to be "Ready" ...
I0920 20:49:16.800356 17609 pod_ready.go:93] pod "kube-apiserver-addons-135472" in "kube-system" namespace has status "Ready":"True"
I0920 20:49:16.800372 17609 pod_ready.go:82] duration metric: took 3.378557ms for pod "kube-apiserver-addons-135472" in "kube-system" namespace to be "Ready" ...
I0920 20:49:16.800379 17609 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-135472" in "kube-system" namespace to be "Ready" ...
I0920 20:49:16.804029 17609 pod_ready.go:93] pod "kube-controller-manager-addons-135472" in "kube-system" namespace has status "Ready":"True"
I0920 20:49:16.804045 17609 pod_ready.go:82] duration metric: took 3.65981ms for pod "kube-controller-manager-addons-135472" in "kube-system" namespace to be "Ready" ...
I0920 20:49:16.804052 17609 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-dldq9" in "kube-system" namespace to be "Ready" ...
I0920 20:49:16.807463 17609 pod_ready.go:93] pod "kube-proxy-dldq9" in "kube-system" namespace has status "Ready":"True"
I0920 20:49:16.807486 17609 pod_ready.go:82] duration metric: took 3.426976ms for pod "kube-proxy-dldq9" in "kube-system" namespace to be "Ready" ...
I0920 20:49:16.807498 17609 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-135472" in "kube-system" namespace to be "Ready" ...
I0920 20:49:17.074606 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:17.191157 17609 pod_ready.go:93] pod "kube-scheduler-addons-135472" in "kube-system" namespace has status "Ready":"True"
I0920 20:49:17.191179 17609 pod_ready.go:82] duration metric: took 383.668405ms for pod "kube-scheduler-addons-135472" in "kube-system" namespace to be "Ready" ...
I0920 20:49:17.191193 17609 pod_ready.go:39] duration metric: took 40.919342338s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0920 20:49:17.191211 17609 api_server.go:52] waiting for apiserver process to appear ...
I0920 20:49:17.191252 17609 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0920 20:49:17.204500 17609 api_server.go:72] duration metric: took 42.639064821s to wait for apiserver process to appear ...
I0920 20:49:17.204520 17609 api_server.go:88] waiting for apiserver healthz status ...
I0920 20:49:17.204536 17609 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0920 20:49:17.208581 17609 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0920 20:49:17.209260 17609 api_server.go:141] control plane version: v1.31.1
I0920 20:49:17.209279 17609 api_server.go:131] duration metric: took 4.753775ms to wait for apiserver health ...
I0920 20:49:17.209286 17609 system_pods.go:43] waiting for kube-system pods to appear ...
I0920 20:49:17.267206 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:17.397686 17609 system_pods.go:59] 17 kube-system pods found
I0920 20:49:17.397718 17609 system_pods.go:61] "coredns-7c65d6cfc9-7m4lj" [6863697a-ab69-4209-927b-5e01cf8662c7] Running
I0920 20:49:17.397729 17609 system_pods.go:61] "csi-hostpath-attacher-0" [c1e886ed-1aed-4f03-9e05-9e66b3718c86] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0920 20:49:17.397738 17609 system_pods.go:61] "csi-hostpath-resizer-0" [2471632c-09f9-46f2-aaef-f5cf1e545204] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0920 20:49:17.397749 17609 system_pods.go:61] "csi-hostpathplugin-g6nz5" [ecd4b7e0-fa89-4066-a44c-68c63ed08848] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0920 20:49:17.397756 17609 system_pods.go:61] "etcd-addons-135472" [68f0a726-c9cf-44a9-bf2d-7b7f27e16a35] Running
I0920 20:49:17.397762 17609 system_pods.go:61] "kube-apiserver-addons-135472" [94bb49d1-b91f-46b7-a3aa-4feb05bc246c] Running
I0920 20:49:17.397767 17609 system_pods.go:61] "kube-controller-manager-addons-135472" [d96a7e2e-0108-4c99-a8c1-5778b9337615] Running
I0920 20:49:17.397776 17609 system_pods.go:61] "kube-ingress-dns-minikube" [e6ce2cf8-833a-4dd3-8adb-39ec6ff7d3a8] Running
I0920 20:49:17.397783 17609 system_pods.go:61] "kube-proxy-dldq9" [04a9e54c-a3de-4a38-9a6d-3fe04c7a3b0a] Running
I0920 20:49:17.397792 17609 system_pods.go:61] "kube-scheduler-addons-135472" [e0299e25-06ee-4577-8908-e7a8b13114c7] Running
I0920 20:49:17.397800 17609 system_pods.go:61] "metrics-server-84c5f94fbc-6j7n7" [0f128981-52f4-40ca-a230-ae60a04056dd] Running
I0920 20:49:17.397808 17609 system_pods.go:61] "nvidia-device-plugin-daemonset-nc7bn" [1891b171-e5c4-4a39-bf97-52c73162793d] Running
I0920 20:49:17.397813 17609 system_pods.go:61] "registry-66c9cd494c-n8x7q" [ab1423e5-b667-4a7f-96f5-061bb4596eeb] Running
I0920 20:49:17.397830 17609 system_pods.go:61] "registry-proxy-8z8jc" [db35f6da-74a1-46c1-8ca2-4c7e51bf1986] Running
I0920 20:49:17.397841 17609 system_pods.go:61] "snapshot-controller-56fcc65765-fn8md" [edd76721-3199-440d-bd9a-5b1ab6bd27e9] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 20:49:17.397849 17609 system_pods.go:61] "snapshot-controller-56fcc65765-zrr94" [ec3d66ea-b501-4ed0-8a23-94005c56a42e] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 20:49:17.397859 17609 system_pods.go:61] "storage-provisioner" [610990bc-6df6-45dc-99ee-7471b9ecd2dc] Running
I0920 20:49:17.397866 17609 system_pods.go:74] duration metric: took 188.574224ms to wait for pod list to return data ...
I0920 20:49:17.397875 17609 default_sa.go:34] waiting for default service account to be created ...
I0920 20:49:17.576129 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:17.591064 17609 default_sa.go:45] found service account: "default"
I0920 20:49:17.591085 17609 default_sa.go:55] duration metric: took 193.203256ms for default service account to be created ...
I0920 20:49:17.591096 17609 system_pods.go:116] waiting for k8s-apps to be running ...
I0920 20:49:17.766801 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:17.820039 17609 system_pods.go:86] 17 kube-system pods found
I0920 20:49:17.820074 17609 system_pods.go:89] "coredns-7c65d6cfc9-7m4lj" [6863697a-ab69-4209-927b-5e01cf8662c7] Running
I0920 20:49:17.820085 17609 system_pods.go:89] "csi-hostpath-attacher-0" [c1e886ed-1aed-4f03-9e05-9e66b3718c86] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0920 20:49:17.820091 17609 system_pods.go:89] "csi-hostpath-resizer-0" [2471632c-09f9-46f2-aaef-f5cf1e545204] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0920 20:49:17.820099 17609 system_pods.go:89] "csi-hostpathplugin-g6nz5" [ecd4b7e0-fa89-4066-a44c-68c63ed08848] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0920 20:49:17.820104 17609 system_pods.go:89] "etcd-addons-135472" [68f0a726-c9cf-44a9-bf2d-7b7f27e16a35] Running
I0920 20:49:17.820109 17609 system_pods.go:89] "kube-apiserver-addons-135472" [94bb49d1-b91f-46b7-a3aa-4feb05bc246c] Running
I0920 20:49:17.820113 17609 system_pods.go:89] "kube-controller-manager-addons-135472" [d96a7e2e-0108-4c99-a8c1-5778b9337615] Running
I0920 20:49:17.820119 17609 system_pods.go:89] "kube-ingress-dns-minikube" [e6ce2cf8-833a-4dd3-8adb-39ec6ff7d3a8] Running
I0920 20:49:17.820123 17609 system_pods.go:89] "kube-proxy-dldq9" [04a9e54c-a3de-4a38-9a6d-3fe04c7a3b0a] Running
I0920 20:49:17.820129 17609 system_pods.go:89] "kube-scheduler-addons-135472" [e0299e25-06ee-4577-8908-e7a8b13114c7] Running
I0920 20:49:17.820135 17609 system_pods.go:89] "metrics-server-84c5f94fbc-6j7n7" [0f128981-52f4-40ca-a230-ae60a04056dd] Running
I0920 20:49:17.820140 17609 system_pods.go:89] "nvidia-device-plugin-daemonset-nc7bn" [1891b171-e5c4-4a39-bf97-52c73162793d] Running
I0920 20:49:17.820146 17609 system_pods.go:89] "registry-66c9cd494c-n8x7q" [ab1423e5-b667-4a7f-96f5-061bb4596eeb] Running
I0920 20:49:17.820150 17609 system_pods.go:89] "registry-proxy-8z8jc" [db35f6da-74a1-46c1-8ca2-4c7e51bf1986] Running
I0920 20:49:17.820159 17609 system_pods.go:89] "snapshot-controller-56fcc65765-fn8md" [edd76721-3199-440d-bd9a-5b1ab6bd27e9] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 20:49:17.820171 17609 system_pods.go:89] "snapshot-controller-56fcc65765-zrr94" [ec3d66ea-b501-4ed0-8a23-94005c56a42e] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 20:49:17.820177 17609 system_pods.go:89] "storage-provisioner" [610990bc-6df6-45dc-99ee-7471b9ecd2dc] Running
I0920 20:49:17.820187 17609 system_pods.go:126] duration metric: took 229.085318ms to wait for k8s-apps to be running ...
I0920 20:49:17.820196 17609 system_svc.go:44] waiting for kubelet service to be running ....
I0920 20:49:17.820239 17609 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0920 20:49:17.830815 17609 system_svc.go:56] duration metric: took 10.611067ms WaitForService to wait for kubelet
I0920 20:49:17.830840 17609 kubeadm.go:582] duration metric: took 43.265404167s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0920 20:49:17.830861 17609 node_conditions.go:102] verifying NodePressure condition ...
I0920 20:49:17.992021 17609 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0920 20:49:17.992050 17609 node_conditions.go:123] node cpu capacity is 8
I0920 20:49:17.992063 17609 node_conditions.go:105] duration metric: took 161.197221ms to run NodePressure ...
I0920 20:49:17.992078 17609 start.go:241] waiting for startup goroutines ...
I0920 20:49:18.074955 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:18.267568 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:18.575707 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:18.766561 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:19.123896 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:19.267155 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:19.575562 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:19.766143 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:20.076020 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:20.266570 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:20.574969 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:20.766751 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:21.077222 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:21.267356 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:21.576063 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:21.766903 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:22.074573 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:22.266551 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:22.575986 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:22.766600 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:23.075162 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:23.267023 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:23.575784 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:23.766914 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:24.077671 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:24.266590 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:24.575860 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:24.766881 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:25.075931 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:25.267428 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:25.575596 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:25.767846 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:26.078315 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:26.266370 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:26.574730 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:26.765974 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:27.075562 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:27.266853 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:27.575449 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:27.766570 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:28.075540 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:28.267032 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:28.576040 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 20:49:28.766136 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:29.075375 17609 kapi.go:107] duration metric: took 41.504002981s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0920 20:49:29.266299 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:29.765934 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:30.265813 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:30.766527 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:31.266311 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:31.766746 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:32.266507 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:32.766153 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:33.266211 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:33.766498 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:34.266219 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:34.766532 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:35.265827 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:35.766472 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:36.265950 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:36.766395 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:37.266246 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:37.766414 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:38.265587 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:38.766044 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:39.266340 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:39.766163 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:40.266140 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:40.766461 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:41.265912 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:41.766604 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:42.267056 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:42.766395 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:43.266707 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:43.766218 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:44.266544 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:44.766342 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:45.266301 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:45.766380 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:46.266575 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:46.766321 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:47.266154 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:47.766681 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:48.266924 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:48.767354 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:49.267018 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:49.766703 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:50.266938 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:50.767742 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:51.267605 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:51.767035 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:52.266538 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:52.766497 17609 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 20:49:53.266510 17609 kapi.go:107] duration metric: took 1m7.503740199s to wait for app.kubernetes.io/name=ingress-nginx ...
I0920 20:50:12.404405 17609 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0920 20:50:12.404429 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:12.905384 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:13.403940 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:13.904268 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:14.404679 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:14.904750 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:15.403988 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:15.904983 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:16.404924 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:16.904634 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:17.403611 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:17.903735 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:18.404201 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:18.904625 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:19.403950 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:19.905585 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:20.403874 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:20.904139 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:21.405508 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:21.903752 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:22.404328 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:22.905014 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:23.404436 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:23.904100 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:24.404858 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:24.904832 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:25.404370 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:25.903894 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:26.404645 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:26.903960 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:27.404342 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:27.905046 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:28.406252 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:28.903944 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:29.404535 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:29.904106 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:30.404694 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:30.904193 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:31.405032 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:31.904511 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:32.404071 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:32.904858 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:33.404477 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:33.904207 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:34.405057 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:34.904119 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:35.404754 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:35.903725 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:36.404054 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:36.904551 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:37.403833 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:37.904382 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:38.404899 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:38.904573 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:39.403981 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:39.905291 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:40.405008 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:40.904892 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:41.404591 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:41.903907 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:42.404350 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:42.904099 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:43.405149 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:43.904783 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:44.403881 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:44.904973 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:45.404535 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:45.904308 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:46.404977 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:46.904410 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:47.404604 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:47.903913 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:48.404448 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:48.904023 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:49.405768 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:49.904742 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:50.404024 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:50.904861 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:51.404344 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:51.904054 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:52.404559 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:52.904561 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:53.404528 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:53.903788 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:54.403978 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:54.904523 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:55.404246 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:55.904402 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:56.403883 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:56.904705 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:57.403951 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:57.904220 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:58.404888 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:58.904593 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:59.403915 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:50:59.904502 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:00.404057 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:00.904747 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:01.404389 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:01.903815 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:02.404755 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:02.904126 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:03.404984 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:03.904389 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:04.403934 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:04.905087 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:05.404766 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:05.904568 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:06.404856 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:06.903655 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:07.404295 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:07.904476 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:08.404088 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:08.904620 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:09.403888 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:09.905014 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:10.404591 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:10.904085 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:11.404888 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:11.904285 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:12.403843 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:12.903931 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:13.404477 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:13.904131 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:14.404769 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:14.904671 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:15.403989 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:15.904854 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:16.404722 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:16.904452 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:17.405172 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:17.905318 17609 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 20:51:18.403845 17609 kapi.go:107] duration metric: took 2m29.502411524s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0920 20:51:18.405198 17609 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-135472 cluster.
I0920 20:51:18.406500 17609 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0920 20:51:18.407689 17609 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0920 20:51:18.409009 17609 out.go:177] * Enabled addons: nvidia-device-plugin, volcano, ingress-dns, cloud-spanner, storage-provisioner, metrics-server, inspektor-gadget, yakd, default-storageclass, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0920 20:51:18.410236 17609 addons.go:510] duration metric: took 2m43.844753984s for enable addons: enabled=[nvidia-device-plugin volcano ingress-dns cloud-spanner storage-provisioner metrics-server inspektor-gadget yakd default-storageclass volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0920 20:51:18.410273 17609 start.go:246] waiting for cluster config update ...
I0920 20:51:18.410297 17609 start.go:255] writing updated cluster config ...
I0920 20:51:18.410542 17609 ssh_runner.go:195] Run: rm -f paused
I0920 20:51:18.457244 17609 start.go:600] kubectl: 1.31.1, cluster: 1.31.1 (minor skew: 0)
I0920 20:51:18.458650 17609 out.go:177] * Done! kubectl is now configured to use "addons-135472" cluster and "default" namespace by default
==> Docker <==
Sep 20 21:00:40 addons-135472 dockerd[1342]: time="2024-09-20T21:00:40.082316704Z" level=info msg="ignoring event" container=3416c221be5ffe8e72a88e8b9e0650f5ca11fc02e4fa48190fba3e0bc9c874fa module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:40 addons-135472 dockerd[1342]: time="2024-09-20T21:00:40.082370825Z" level=info msg="ignoring event" container=3dd570a20fa88d6b7085cad1e4bdd10fc57d98cdf3d23929a8a61ff249d7e70e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:40 addons-135472 dockerd[1342]: time="2024-09-20T21:00:40.160707821Z" level=info msg="ignoring event" container=d3b6af70bf62eb597596b0b8ca0f4a71cfd8f28edf33e8c124975419c2f25398 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:40 addons-135472 dockerd[1342]: time="2024-09-20T21:00:40.169738192Z" level=info msg="ignoring event" container=44a8cbf55922ff942882512e51f4c7b452e8bac6f37678ee5d93cf570b5e1523 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:40 addons-135472 dockerd[1342]: time="2024-09-20T21:00:40.170741349Z" level=info msg="ignoring event" container=82a8eb43f0a917f38f7cae504a87c4bb24bf8666ad72291f25f02afe2c47a798 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:40 addons-135472 dockerd[1342]: time="2024-09-20T21:00:40.173061067Z" level=info msg="ignoring event" container=05a3b1bf7827e9832f874b7165cd7546fc4fca2b3e5f64b7a1a26d78011a59b2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:40 addons-135472 dockerd[1342]: time="2024-09-20T21:00:40.184077267Z" level=info msg="ignoring event" container=9ef7aa4bdb1d976420a088e62ed425779e6b7b97fb8fd7d4951a764048643c4a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:40 addons-135472 dockerd[1342]: time="2024-09-20T21:00:40.791492498Z" level=info msg="ignoring event" container=7f00d1eef828c05468b79ce16832617bdf71250d445733b2857790a0d37e3e70 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:40 addons-135472 dockerd[1342]: time="2024-09-20T21:00:40.805745811Z" level=info msg="ignoring event" container=091c613362bf17cd23ff605967987c32c3148fdcfd3d214f085396f97bcc6a57 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:40 addons-135472 dockerd[1342]: time="2024-09-20T21:00:40.875650571Z" level=info msg="ignoring event" container=6709086385f0eff837a09e42c4738f5b20b434a61ddb44faf997107366d046c1 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:44 addons-135472 dockerd[1342]: time="2024-09-20T21:00:44.094842200Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=d5aa45762762da292a19bb27e5536921054810a397ad6cec4fa0e74b92c8b2d9 spanID=42c0330ef9ba6afa traceID=40bf797b57d65f34a10dee9ca7d9c8fa
Sep 20 21:00:44 addons-135472 dockerd[1342]: time="2024-09-20T21:00:44.155342794Z" level=info msg="ignoring event" container=d5aa45762762da292a19bb27e5536921054810a397ad6cec4fa0e74b92c8b2d9 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:44 addons-135472 cri-dockerd[1608]: time="2024-09-20T21:00:44Z" level=info msg="Failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod \"ingress-nginx-controller-bc57996ff-s8phk_ingress-nginx\": unexpected command output Device \"eth0\" does not exist.\n with error: exit status 1"
Sep 20 21:00:44 addons-135472 dockerd[1342]: time="2024-09-20T21:00:44.299536103Z" level=info msg="ignoring event" container=748b52dd51b500d4f7c30d6d13da3b101f2adaed5cca0c9dee9a6dd57dc7bf85 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:46 addons-135472 dockerd[1342]: time="2024-09-20T21:00:46.371481499Z" level=info msg="ignoring event" container=a66b2384155c92184f0a4508370d03e8fd8a539515d0c688c129a0c9cb52144a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:46 addons-135472 dockerd[1342]: time="2024-09-20T21:00:46.372821446Z" level=info msg="ignoring event" container=07c696666a296d6c40c9b8e1991e653d6bad920ad15d2e09cfdfed969cfe9560 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:46 addons-135472 dockerd[1342]: time="2024-09-20T21:00:46.547456284Z" level=info msg="ignoring event" container=ffe65bfbcb0ef9f723a71a57dbec76888af6efdd81dbf4b65a5480fbbbccc2db module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:46 addons-135472 dockerd[1342]: time="2024-09-20T21:00:46.569313545Z" level=info msg="ignoring event" container=234433d87febf06e7d2df942b3c647f500b842f1f86e6b03d913cb50c642aac0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:00:49 addons-135472 dockerd[1342]: time="2024-09-20T21:00:49.846450557Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed" spanID=cc40d4310e127e80 traceID=32183bc587e51af6903141c7808b55e4
Sep 20 21:00:49 addons-135472 dockerd[1342]: time="2024-09-20T21:00:49.848068722Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed" spanID=cc40d4310e127e80 traceID=32183bc587e51af6903141c7808b55e4
Sep 20 21:01:09 addons-135472 dockerd[1342]: time="2024-09-20T21:01:09.643581076Z" level=info msg="ignoring event" container=6641c319a32596647247533456772d6a0d4089f23b7aab8dbde2342175a4a136 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:01:10 addons-135472 dockerd[1342]: time="2024-09-20T21:01:10.086896514Z" level=info msg="ignoring event" container=24b3d0c671515d361568eaad383b4db480e7769618191129552d7c3bfad73ed0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:01:10 addons-135472 dockerd[1342]: time="2024-09-20T21:01:10.171136172Z" level=info msg="ignoring event" container=0041c6fb54cc137712d2058d0fe6c716d26f9ac789549b6f9d7bbc62fc1134fd module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:01:10 addons-135472 dockerd[1342]: time="2024-09-20T21:01:10.224939270Z" level=info msg="ignoring event" container=ab9e92e6feaa1cf4fec9ce691b379d0b46c7258d4210549832c58437f61b19e9 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 21:01:10 addons-135472 dockerd[1342]: time="2024-09-20T21:01:10.313408675Z" level=info msg="ignoring event" container=59a42f110ab6128e8493a4ed4537ceb939f6fcda3ee046cfdec51700e21f9685 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
2da9301ea32c7 kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 31 seconds ago Running hello-world-app 0 6e2065431b2ab hello-world-app-55bf9c44b4-6gd9l
17a7916bd4793 nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf 38 seconds ago Running nginx 0 b5b4651ace8eb nginx
ef3621a95b76a gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 3b06f3f4cf97f gcp-auth-89d5ffd79-89jmh
2614224e15df0 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited patch 0 a00185430995f ingress-nginx-admission-patch-bdkdp
b4dab13c6b5be registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 4f9dc6e057bef ingress-nginx-admission-create-72gkj
0f82d95cccb42 rancher/local-path-provisioner@sha256:e34c88ae0affb1cdefbb874140d6339d4a27ec4ee420ae8199cd839997b05246 12 minutes ago Running local-path-provisioner 0 a6c0e047973fe local-path-provisioner-86d989889c-mj2z7
1a277a4906625 6e38f40d628db 12 minutes ago Running storage-provisioner 0 bd433ed5b94b2 storage-provisioner
d4549c22b98ac c69fa2e9cbf5f 12 minutes ago Running coredns 0 f906d9e567ff7 coredns-7c65d6cfc9-7m4lj
0d0804c9fcbff 60c005f310ff3 12 minutes ago Running kube-proxy 0 3c52512d3c308 kube-proxy-dldq9
f85fc9ae7c205 6bab7719df100 12 minutes ago Running kube-apiserver 0 c1d70a89071f9 kube-apiserver-addons-135472
0817a42f2a24e 9aa1fad941575 12 minutes ago Running kube-scheduler 0 dd09603121160 kube-scheduler-addons-135472
d228da92f9f07 2e96e5913fc06 12 minutes ago Running etcd 0 102b3750a64c9 etcd-addons-135472
cdff8f7c2b923 175ffd71cce3d 12 minutes ago Running kube-controller-manager 0 9e61c84a88451 kube-controller-manager-addons-135472
==> coredns [d4549c22b98a] <==
[INFO] 10.244.0.21:32891 - 51378 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00594696s
[INFO] 10.244.0.21:44163 - 30159 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000069706s
[INFO] 10.244.0.21:34449 - 26402 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.010218436s
[INFO] 10.244.0.21:56418 - 50882 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.001568911s
[INFO] 10.244.0.21:38275 - 2755 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.002950694s
[INFO] 10.244.0.21:41210 - 16321 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.010397775s
[INFO] 10.244.0.21:32891 - 48383 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.002718676s
[INFO] 10.244.0.21:36523 - 42172 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.010771909s
[INFO] 10.244.0.21:41781 - 53445 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00299836s
[INFO] 10.244.0.21:38865 - 9563 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003069156s
[INFO] 10.244.0.21:56446 - 966 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.010993413s
[INFO] 10.244.0.21:34449 - 7986 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000243639s
[INFO] 10.244.0.21:36523 - 60842 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000078574s
[INFO] 10.244.0.21:32891 - 3257 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000050826s
[INFO] 10.244.0.21:38057 - 19560 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.002638223s
[INFO] 10.244.0.21:56446 - 63068 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000064117s
[INFO] 10.244.0.21:39362 - 17236 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004764669s
[INFO] 10.244.0.21:41781 - 1814 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000082504s
[INFO] 10.244.0.21:38275 - 62761 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00004945s
[INFO] 10.244.0.21:38865 - 64813 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000059476s
[INFO] 10.244.0.21:56418 - 29113 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000044093s
[INFO] 10.244.0.21:41210 - 19662 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000195825s
[INFO] 10.244.0.21:38057 - 58879 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000053766s
[INFO] 10.244.0.21:39362 - 23095 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00475268s
[INFO] 10.244.0.21:39362 - 25605 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00008046s
==> describe nodes <==
Name: addons-135472
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-135472
kubernetes.io/os=linux
minikube.k8s.io/commit=b921bee7dddd4990dd76a4773b23d7ec11e6144f
minikube.k8s.io/name=addons-135472
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_20T20_48_30_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-135472
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Fri, 20 Sep 2024 20:48:27 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-135472
AcquireTime: <unset>
RenewTime: Fri, 20 Sep 2024 21:01:04 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Fri, 20 Sep 2024 21:01:06 +0000 Fri, 20 Sep 2024 20:48:26 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Fri, 20 Sep 2024 21:01:06 +0000 Fri, 20 Sep 2024 20:48:26 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Fri, 20 Sep 2024 21:01:06 +0000 Fri, 20 Sep 2024 20:48:26 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Fri, 20 Sep 2024 21:01:06 +0000 Fri, 20 Sep 2024 20:48:27 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-135472
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859320Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859320Ki
pods: 110
System Info:
Machine ID: 94a5a244c6f24084b87142a7c0652553
System UUID: 63f110a0-0a70-4a3b-8138-a16e3d9aa477
Boot ID: f541ecf7-517e-485c-8b68-8f94d94b6d3f
Kernel Version: 5.15.0-1069-gcp
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.3.0
Kubelet Version: v1.31.1
Kube-Proxy Version: v1.31.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (12 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m15s
default hello-world-app-55bf9c44b4-6gd9l 0 (0%) 0 (0%) 0 (0%) 0 (0%) 33s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 41s
gcp-auth gcp-auth-89d5ffd79-89jmh 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
kube-system coredns-7c65d6cfc9-7m4lj 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-135472 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-135472 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-135472 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-dldq9 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-135472 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
local-path-storage local-path-provisioner-86d989889c-mj2z7 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeHasSufficientMemory 12m (x8 over 12m) kubelet Node addons-135472 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m (x7 over 12m) kubelet Node addons-135472 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m (x7 over 12m) kubelet Node addons-135472 status is now: NodeHasSufficientPID
Normal Starting 12m kubelet Starting kubelet.
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-135472 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-135472 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-135472 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-135472 event: Registered Node addons-135472 in Controller
==> dmesg <==
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 9e 70 77 6b 65 91 08 06
[ +1.633603] IPv4: martian source 10.244.0.1 from 10.244.0.17, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff e2 b6 b7 08 60 16 08 06
[ +5.135270] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 46 4a dd 84 01 4d 08 06
[ +0.590819] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 56 49 1b bc 81 d2 08 06
[ +0.088986] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff f6 4a 9f fa 69 cd 08 06
[ +25.450389] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 82 cf 16 3a 98 6a 08 06
[ +1.010954] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff f2 f0 b4 01 c9 ac 08 06
[Sep20 20:50] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff ea 74 43 84 da ff 08 06
[ +0.009410] IPv4: martian source 10.244.0.1 from 10.244.0.23, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff fe e5 20 aa ce 99 08 06
[Sep20 20:51] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff f2 72 f2 79 4b 13 08 06
[ +0.000491] IPv4: martian source 10.244.0.25 from 10.244.0.3, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff 2e 31 fd 29 84 28 08 06
[Sep20 21:00] IPv4: martian source 10.244.0.35 from 10.244.0.21, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 82 cf 16 3a 98 6a 08 06
[ +1.858654] IPv4: martian source 10.244.0.21 from 10.244.0.3, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 2e 31 fd 29 84 28 08 06
==> etcd [d228da92f9f0] <==
{"level":"info","ts":"2024-09-20T20:48:25.584117Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-09-20T20:48:25.584131Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-20T20:48:25.584956Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T20:48:25.585602Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-135472 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-09-20T20:48:25.585607Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-20T20:48:25.585634Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-20T20:48:25.585840Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-20T20:48:25.585869Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-20T20:48:25.586319Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T20:48:25.586385Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T20:48:25.586405Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T20:48:25.586767Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-20T20:48:25.587035Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-20T20:48:25.587925Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-09-20T20:48:25.588092Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-20T20:49:10.631392Z","caller":"traceutil/trace.go:171","msg":"trace[1058944811] transaction","detail":"{read_only:false; response_revision:1068; number_of_response:1; }","duration":"123.049172ms","start":"2024-09-20T20:49:10.508323Z","end":"2024-09-20T20:49:10.631372Z","steps":["trace[1058944811] 'process raft request' (duration: 56.658154ms)","trace[1058944811] 'compare' (duration: 66.318188ms)"],"step_count":2}
{"level":"warn","ts":"2024-09-20T20:49:12.678779Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"102.695085ms","expected-duration":"100ms","prefix":"read-only range ","request":"limit:1 keys_only:true ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-20T20:49:12.678873Z","caller":"traceutil/trace.go:171","msg":"trace[1110733045] range","detail":"{range_begin:; range_end:; response_count:0; response_revision:1075; }","duration":"102.806898ms","start":"2024-09-20T20:49:12.576053Z","end":"2024-09-20T20:49:12.678860Z","steps":["trace[1110733045] 'range keys from in-memory index tree' (duration: 102.655162ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-20T20:49:12.678758Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"105.921167ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-20T20:49:12.678965Z","caller":"traceutil/trace.go:171","msg":"trace[1118587257] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:1075; }","duration":"106.164017ms","start":"2024-09-20T20:49:12.572787Z","end":"2024-09-20T20:49:12.678951Z","steps":["trace[1118587257] 'range keys from in-memory index tree' (duration: 105.885963ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-20T20:49:52.088930Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"115.121626ms","expected-duration":"100ms","prefix":"","request":"header:<ID:8128032030038373407 username:\"kube-apiserver-etcd-client\" auth_revision:1 > txn:<compare:<target:MOD key:\"/registry/events/gcp-auth/gcp-auth-certs-create.17f70ede062708ce\" mod_revision:1141 > success:<request_put:<key:\"/registry/events/gcp-auth/gcp-auth-certs-create.17f70ede062708ce\" value_size:819 lease:8128032030038373060 >> failure:<request_range:<key:\"/registry/events/gcp-auth/gcp-auth-certs-create.17f70ede062708ce\" > >>","response":"size:16"}
{"level":"info","ts":"2024-09-20T20:49:52.089020Z","caller":"traceutil/trace.go:171","msg":"trace[33980969] transaction","detail":"{read_only:false; response_revision:1270; number_of_response:1; }","duration":"168.922468ms","start":"2024-09-20T20:49:51.920083Z","end":"2024-09-20T20:49:52.089005Z","steps":["trace[33980969] 'process raft request' (duration: 53.276544ms)","trace[33980969] 'compare' (duration: 115.042022ms)"],"step_count":2}
{"level":"info","ts":"2024-09-20T20:58:26.578734Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1895}
{"level":"info","ts":"2024-09-20T20:58:26.601591Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1895,"took":"22.37242ms","hash":785413254,"current-db-size-bytes":8921088,"current-db-size":"8.9 MB","current-db-size-in-use-bytes":4964352,"current-db-size-in-use":"5.0 MB"}
{"level":"info","ts":"2024-09-20T20:58:26.601625Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":785413254,"revision":1895,"compact-revision":-1}
==> gcp-auth [ef3621a95b76] <==
2024/09/20 20:51:56 Ready to write response ...
2024/09/20 20:51:56 Ready to marshal response ...
2024/09/20 20:51:56 Ready to write response ...
2024/09/20 21:00:09 Ready to marshal response ...
2024/09/20 21:00:09 Ready to write response ...
2024/09/20 21:00:09 Ready to marshal response ...
2024/09/20 21:00:09 Ready to write response ...
2024/09/20 21:00:09 Ready to marshal response ...
2024/09/20 21:00:09 Ready to write response ...
2024/09/20 21:00:11 Ready to marshal response ...
2024/09/20 21:00:11 Ready to write response ...
2024/09/20 21:00:15 Ready to marshal response ...
2024/09/20 21:00:15 Ready to write response ...
2024/09/20 21:00:15 Ready to marshal response ...
2024/09/20 21:00:15 Ready to write response ...
2024/09/20 21:00:15 Ready to marshal response ...
2024/09/20 21:00:15 Ready to write response ...
2024/09/20 21:00:18 Ready to marshal response ...
2024/09/20 21:00:18 Ready to write response ...
2024/09/20 21:00:29 Ready to marshal response ...
2024/09/20 21:00:29 Ready to write response ...
2024/09/20 21:00:30 Ready to marshal response ...
2024/09/20 21:00:30 Ready to write response ...
2024/09/20 21:00:38 Ready to marshal response ...
2024/09/20 21:00:38 Ready to write response ...
==> kernel <==
21:01:11 up 43 min, 0 users, load average: 0.54, 0.33, 0.27
Linux addons-135472 5.15.0-1069-gcp #77~20.04.1-Ubuntu SMP Sun Sep 1 19:39:16 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [f85fc9ae7c20] <==
W0920 20:51:47.973469 1 cacher.go:171] Terminating all watchers from cacher queues.scheduling.volcano.sh
W0920 20:51:48.076606 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0920 20:51:48.275507 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0920 20:51:48.585002 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0920 21:00:08.990734 1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
I0920 21:00:15.675670 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.99.43.158"}
I0920 21:00:18.805839 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0920 21:00:24.475924 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0920 21:00:25.490955 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0920 21:00:29.917946 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0920 21:00:30.265919 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.100.155.243"}
I0920 21:00:38.731790 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.103.173.68"}
I0920 21:00:46.228880 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0920 21:00:46.228934 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0920 21:00:46.239965 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0920 21:00:46.240011 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0920 21:00:46.242848 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0920 21:00:46.242892 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0920 21:00:46.260323 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0920 21:00:46.260371 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0920 21:00:46.269172 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0920 21:00:46.269204 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0920 21:00:47.243666 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W0920 21:00:47.269717 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
W0920 21:00:47.279403 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
==> kube-controller-manager [cdff8f7c2b92] <==
E0920 21:00:52.768105 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 21:00:53.786206 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 21:00:53.786240 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 21:00:57.079939 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 21:00:57.079979 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 21:00:57.322383 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 21:00:57.322416 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 21:00:57.533621 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 21:00:57.533658 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 21:01:04.379028 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 21:01:04.379065 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 21:01:04.629308 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 21:01:04.629346 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0920 21:01:04.956401 1 shared_informer.go:313] Waiting for caches to sync for resource quota
I0920 21:01:04.956431 1 shared_informer.go:320] Caches are synced for resource quota
I0920 21:01:05.263208 1 shared_informer.go:313] Waiting for caches to sync for garbage collector
I0920 21:01:05.263250 1 shared_informer.go:320] Caches are synced for garbage collector
W0920 21:01:06.107708 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 21:01:06.107745 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0920 21:01:06.405383 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-135472"
W0920 21:01:06.993320 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 21:01:06.993359 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 21:01:08.734381 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 21:01:08.734417 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0920 21:01:10.054970 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="8.322µs"
==> kube-proxy [0d0804c9fcbf] <==
I0920 20:48:38.575560 1 server_linux.go:66] "Using iptables proxy"
I0920 20:48:39.172955 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0920 20:48:39.173022 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0920 20:48:39.757807 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0920 20:48:39.757865 1 server_linux.go:169] "Using iptables Proxier"
I0920 20:48:39.766777 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0920 20:48:39.767193 1 server.go:483] "Version info" version="v1.31.1"
I0920 20:48:39.767217 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0920 20:48:39.769542 1 config.go:199] "Starting service config controller"
I0920 20:48:39.769566 1 shared_informer.go:313] Waiting for caches to sync for service config
I0920 20:48:39.769590 1 config.go:105] "Starting endpoint slice config controller"
I0920 20:48:39.769596 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0920 20:48:39.770003 1 config.go:328] "Starting node config controller"
I0920 20:48:39.770009 1 shared_informer.go:313] Waiting for caches to sync for node config
I0920 20:48:39.870467 1 shared_informer.go:320] Caches are synced for node config
I0920 20:48:39.870504 1 shared_informer.go:320] Caches are synced for service config
I0920 20:48:39.870541 1 shared_informer.go:320] Caches are synced for endpoint slice config
==> kube-scheduler [0817a42f2a24] <==
E0920 20:48:27.671246 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
E0920 20:48:27.671442 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0920 20:48:27.671398 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0920 20:48:27.671449 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 20:48:27.671253 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0920 20:48:27.671589 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 20:48:27.671253 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0920 20:48:27.671632 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0920 20:48:27.671401 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0920 20:48:27.671673 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0920 20:48:28.483651 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0920 20:48:28.483682 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 20:48:28.495740 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0920 20:48:28.495764 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 20:48:28.567223 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0920 20:48:28.567259 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0920 20:48:28.572523 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0920 20:48:28.572558 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0920 20:48:28.580719 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0920 20:48:28.580750 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0920 20:48:28.729016 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0920 20:48:28.729055 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 20:48:28.730008 1 reflector.go:561] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0920 20:48:28.730039 1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
I0920 20:48:31.468803 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 20 21:00:54 addons-135472 kubelet[2444]: E0920 21:00:54.792438 2444 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="837e41f6-532d-44bd-9cbf-429c7ea7bd7d"
Sep 20 21:01:02 addons-135472 kubelet[2444]: I0920 21:01:02.790936 2444 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="kube-system/registry-proxy-8z8jc" secret="" err="secret \"gcp-auth\" not found"
Sep 20 21:01:02 addons-135472 kubelet[2444]: E0920 21:01:02.792762 2444 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox\\\"\"" pod="default/registry-test" podUID="60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c"
Sep 20 21:01:05 addons-135472 kubelet[2444]: I0920 21:01:05.790663 2444 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="kube-system/coredns-7c65d6cfc9-7m4lj" secret="" err="secret \"gcp-auth\" not found"
Sep 20 21:01:06 addons-135472 kubelet[2444]: E0920 21:01:06.792331 2444 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="837e41f6-532d-44bd-9cbf-429c7ea7bd7d"
Sep 20 21:01:09 addons-135472 kubelet[2444]: I0920 21:01:09.786246 2444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mnsc\" (UniqueName: \"kubernetes.io/projected/60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c-kube-api-access-8mnsc\") pod \"60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c\" (UID: \"60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c\") "
Sep 20 21:01:09 addons-135472 kubelet[2444]: I0920 21:01:09.786280 2444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c-gcp-creds\") pod \"60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c\" (UID: \"60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c\") "
Sep 20 21:01:09 addons-135472 kubelet[2444]: I0920 21:01:09.786334 2444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c" (UID: "60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Sep 20 21:01:09 addons-135472 kubelet[2444]: I0920 21:01:09.787811 2444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c-kube-api-access-8mnsc" (OuterVolumeSpecName: "kube-api-access-8mnsc") pod "60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c" (UID: "60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c"). InnerVolumeSpecName "kube-api-access-8mnsc". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 20 21:01:09 addons-135472 kubelet[2444]: I0920 21:01:09.886862 2444 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-8mnsc\" (UniqueName: \"kubernetes.io/projected/60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c-kube-api-access-8mnsc\") on node \"addons-135472\" DevicePath \"\""
Sep 20 21:01:09 addons-135472 kubelet[2444]: I0920 21:01:09.886896 2444 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/60dd3ab7-ff5f-4ccc-8fef-a1694f5efc9c-gcp-creds\") on node \"addons-135472\" DevicePath \"\""
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.390575 2444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn7ch\" (UniqueName: \"kubernetes.io/projected/ab1423e5-b667-4a7f-96f5-061bb4596eeb-kube-api-access-wn7ch\") pod \"ab1423e5-b667-4a7f-96f5-061bb4596eeb\" (UID: \"ab1423e5-b667-4a7f-96f5-061bb4596eeb\") "
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.390636 2444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw2r9\" (UniqueName: \"kubernetes.io/projected/db35f6da-74a1-46c1-8ca2-4c7e51bf1986-kube-api-access-vw2r9\") pod \"db35f6da-74a1-46c1-8ca2-4c7e51bf1986\" (UID: \"db35f6da-74a1-46c1-8ca2-4c7e51bf1986\") "
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.392350 2444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab1423e5-b667-4a7f-96f5-061bb4596eeb-kube-api-access-wn7ch" (OuterVolumeSpecName: "kube-api-access-wn7ch") pod "ab1423e5-b667-4a7f-96f5-061bb4596eeb" (UID: "ab1423e5-b667-4a7f-96f5-061bb4596eeb"). InnerVolumeSpecName "kube-api-access-wn7ch". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.392512 2444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db35f6da-74a1-46c1-8ca2-4c7e51bf1986-kube-api-access-vw2r9" (OuterVolumeSpecName: "kube-api-access-vw2r9") pod "db35f6da-74a1-46c1-8ca2-4c7e51bf1986" (UID: "db35f6da-74a1-46c1-8ca2-4c7e51bf1986"). InnerVolumeSpecName "kube-api-access-vw2r9". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.491129 2444 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-wn7ch\" (UniqueName: \"kubernetes.io/projected/ab1423e5-b667-4a7f-96f5-061bb4596eeb-kube-api-access-wn7ch\") on node \"addons-135472\" DevicePath \"\""
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.491162 2444 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-vw2r9\" (UniqueName: \"kubernetes.io/projected/db35f6da-74a1-46c1-8ca2-4c7e51bf1986-kube-api-access-vw2r9\") on node \"addons-135472\" DevicePath \"\""
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.603714 2444 scope.go:117] "RemoveContainer" containerID="24b3d0c671515d361568eaad383b4db480e7769618191129552d7c3bfad73ed0"
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.620070 2444 scope.go:117] "RemoveContainer" containerID="24b3d0c671515d361568eaad383b4db480e7769618191129552d7c3bfad73ed0"
Sep 20 21:01:10 addons-135472 kubelet[2444]: E0920 21:01:10.620730 2444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 24b3d0c671515d361568eaad383b4db480e7769618191129552d7c3bfad73ed0" containerID="24b3d0c671515d361568eaad383b4db480e7769618191129552d7c3bfad73ed0"
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.620770 2444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"24b3d0c671515d361568eaad383b4db480e7769618191129552d7c3bfad73ed0"} err="failed to get container status \"24b3d0c671515d361568eaad383b4db480e7769618191129552d7c3bfad73ed0\": rpc error: code = Unknown desc = Error response from daemon: No such container: 24b3d0c671515d361568eaad383b4db480e7769618191129552d7c3bfad73ed0"
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.620796 2444 scope.go:117] "RemoveContainer" containerID="0041c6fb54cc137712d2058d0fe6c716d26f9ac789549b6f9d7bbc62fc1134fd"
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.636288 2444 scope.go:117] "RemoveContainer" containerID="0041c6fb54cc137712d2058d0fe6c716d26f9ac789549b6f9d7bbc62fc1134fd"
Sep 20 21:01:10 addons-135472 kubelet[2444]: E0920 21:01:10.636895 2444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 0041c6fb54cc137712d2058d0fe6c716d26f9ac789549b6f9d7bbc62fc1134fd" containerID="0041c6fb54cc137712d2058d0fe6c716d26f9ac789549b6f9d7bbc62fc1134fd"
Sep 20 21:01:10 addons-135472 kubelet[2444]: I0920 21:01:10.637046 2444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"0041c6fb54cc137712d2058d0fe6c716d26f9ac789549b6f9d7bbc62fc1134fd"} err="failed to get container status \"0041c6fb54cc137712d2058d0fe6c716d26f9ac789549b6f9d7bbc62fc1134fd\": rpc error: code = Unknown desc = Error response from daemon: No such container: 0041c6fb54cc137712d2058d0fe6c716d26f9ac789549b6f9d7bbc62fc1134fd"
==> storage-provisioner [1a277a490662] <==
I0920 20:48:41.859972 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0920 20:48:41.883861 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0920 20:48:41.883942 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0920 20:48:41.966756 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0920 20:48:41.966964 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-135472_5f138205-f562-4d76-aeec-c671f8d59364!
I0920 20:48:41.970074 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"6370178b-d8be-414f-9c5a-715141d2a863", APIVersion:"v1", ResourceVersion:"613", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-135472_5f138205-f562-4d76-aeec-c671f8d59364 became leader
I0920 20:48:42.067133 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-135472_5f138205-f562-4d76-aeec-c671f8d59364!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-135472 -n addons-135472
helpers_test.go:261: (dbg) Run: kubectl --context addons-135472 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-135472 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-135472 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-135472/192.168.49.2
Start Time: Fri, 20 Sep 2024 20:51:56 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.27
IPs:
IP: 10.244.0.27
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-npb8r (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-npb8r:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m15s default-scheduler Successfully assigned default/busybox to addons-135472
Normal Pulling 7m56s (x4 over 9m15s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m56s (x4 over 9m15s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m56s (x4 over 9m15s) kubelet Error: ErrImagePull
Warning Failed 7m33s (x6 over 9m14s) kubelet Error: ImagePullBackOff
Normal BackOff 4m3s (x21 over 9m14s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (73.30s)