=== RUN TestAddons/parallel/Ingress
=== PAUSE TestAddons/parallel/Ingress
=== CONT TestAddons/parallel/Ingress
addons_test.go:207: (dbg) Run: kubectl --context addons-109663 wait --for=condition=ready --namespace=ingress-nginx pod --selector=app.kubernetes.io/component=controller --timeout=90s
addons_test.go:232: (dbg) Run: kubectl --context addons-109663 replace --force -f testdata/nginx-ingress-v1.yaml
addons_test.go:245: (dbg) Run: kubectl --context addons-109663 replace --force -f testdata/nginx-pod-svc.yaml
addons_test.go:250: (dbg) TestAddons/parallel/Ingress: waiting 8m0s for pods matching "run=nginx" in namespace "default" ...
helpers_test.go:344: "nginx" [0c78f186-ad36-4191-83a7-fc36688df669] Pending / Ready:ContainersNotReady (containers with unready status: [nginx]) / ContainersReady:ContainersNotReady (containers with unready status: [nginx])
helpers_test.go:344: "nginx" [0c78f186-ad36-4191-83a7-fc36688df669] Running
addons_test.go:250: (dbg) TestAddons/parallel/Ingress: run=nginx healthy within 10.003048741s
I1216 10:35:31.764473 847292 kapi.go:150] Service nginx in namespace default found.
addons_test.go:262: (dbg) Run: out/minikube-linux-amd64 -p addons-109663 ssh "curl -s http://127.0.0.1/ -H 'Host: nginx.example.com'"
addons_test.go:262: (dbg) Non-zero exit: out/minikube-linux-amd64 -p addons-109663 ssh "curl -s http://127.0.0.1/ -H 'Host: nginx.example.com'": exit status 1 (2m10.513370222s)
** stderr **
ssh: Process exited with status 28
** /stderr **
addons_test.go:278: failed to get expected response from http://127.0.0.1/ within minikube: exit status 1
addons_test.go:286: (dbg) Run: kubectl --context addons-109663 replace --force -f testdata/ingress-dns-example-v1.yaml
addons_test.go:291: (dbg) Run: out/minikube-linux-amd64 -p addons-109663 ip
addons_test.go:297: (dbg) Run: nslookup hello-john.test 192.168.49.2
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Ingress]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-109663
helpers_test.go:235: (dbg) docker inspect addons-109663:
-- stdout --
[
{
"Id": "1a5d30b35ebd94d45b4b20f053e9d801bffac6feb46db54b983452fbca50984b",
"Created": "2024-12-16T10:32:42.208735109Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 849348,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-12-16T10:32:42.321849535Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:7036ee4d70b7e266f67949e27a52ed21246dbdde9902b1d29235748548d311cb",
"ResolvConfPath": "/var/lib/docker/containers/1a5d30b35ebd94d45b4b20f053e9d801bffac6feb46db54b983452fbca50984b/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/1a5d30b35ebd94d45b4b20f053e9d801bffac6feb46db54b983452fbca50984b/hostname",
"HostsPath": "/var/lib/docker/containers/1a5d30b35ebd94d45b4b20f053e9d801bffac6feb46db54b983452fbca50984b/hosts",
"LogPath": "/var/lib/docker/containers/1a5d30b35ebd94d45b4b20f053e9d801bffac6feb46db54b983452fbca50984b/1a5d30b35ebd94d45b4b20f053e9d801bffac6feb46db54b983452fbca50984b-json.log",
"Name": "/addons-109663",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-109663:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-109663",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/65804e8ecf53a4a783bcbd11ff1ee57774d652a79d14faa51abcf74021f9f0a6-init/diff:/var/lib/docker/overlay2/123e2f1df366b4ca43a26782c77043f0e4cd5c6388fa90b6b3300da767616189/diff",
"MergedDir": "/var/lib/docker/overlay2/65804e8ecf53a4a783bcbd11ff1ee57774d652a79d14faa51abcf74021f9f0a6/merged",
"UpperDir": "/var/lib/docker/overlay2/65804e8ecf53a4a783bcbd11ff1ee57774d652a79d14faa51abcf74021f9f0a6/diff",
"WorkDir": "/var/lib/docker/overlay2/65804e8ecf53a4a783bcbd11ff1ee57774d652a79d14faa51abcf74021f9f0a6/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-109663",
"Source": "/var/lib/docker/volumes/addons-109663/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-109663",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-109663",
"name.minikube.sigs.k8s.io": "addons-109663",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "6c0a3e7ab167392b4b1457a91806bd78d3a67f0fd8e01a37251db9ff03c74d5d",
"SandboxKey": "/var/run/docker/netns/6c0a3e7ab167",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33139"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33140"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33143"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33141"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33142"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-109663": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "8d8d19425ae9a0d7e09aa1deae754ccc44dc321a7589581cd2cc49ee9d8127e2",
"EndpointID": "1fd1a8fade4259280c934e3bd3078705e00cc2e63230df4f97442f57a51b046a",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-109663",
"1a5d30b35ebd"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-109663 -n addons-109663
helpers_test.go:244: <<< TestAddons/parallel/Ingress FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Ingress]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-109663 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-amd64 -p addons-109663 logs -n 25: (1.09121414s)
helpers_test.go:252: TestAddons/parallel/Ingress logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-only-505735 | download-only-505735 | jenkins | v1.34.0 | 16 Dec 24 10:32 UTC | 16 Dec 24 10:32 UTC |
| start | --download-only -p | download-docker-072674 | jenkins | v1.34.0 | 16 Dec 24 10:32 UTC | |
| | download-docker-072674 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=crio | | | | | |
| delete | -p download-docker-072674 | download-docker-072674 | jenkins | v1.34.0 | 16 Dec 24 10:32 UTC | 16 Dec 24 10:32 UTC |
| start | --download-only -p | binary-mirror-516574 | jenkins | v1.34.0 | 16 Dec 24 10:32 UTC | |
| | binary-mirror-516574 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:32893 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=crio | | | | | |
| delete | -p binary-mirror-516574 | binary-mirror-516574 | jenkins | v1.34.0 | 16 Dec 24 10:32 UTC | 16 Dec 24 10:32 UTC |
| addons | enable dashboard -p | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:32 UTC | |
| | addons-109663 | | | | | |
| addons | disable dashboard -p | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:32 UTC | |
| | addons-109663 | | | | | |
| start | -p addons-109663 --wait=true | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:32 UTC | 16 Dec 24 10:34 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --addons=amd-gpu-device-plugin | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=crio | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| addons | addons-109663 addons disable | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:34 UTC | 16 Dec 24 10:34 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | addons-109663 addons disable | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:34 UTC | 16 Dec 24 10:35 UTC |
| | gcp-auth --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | enable headlamp | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| | -p addons-109663 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-109663 addons | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| | disable nvidia-device-plugin | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-109663 addons disable | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ssh | addons-109663 ssh cat | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| | /opt/local-path-provisioner/pvc-9e504c9a-bb3a-4229-9525-d31715212760_default_test-pvc/file1 | | | | | |
| addons | addons-109663 addons disable | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-109663 ip | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| addons | addons-109663 addons disable | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-109663 addons | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| | disable cloud-spanner | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-109663 addons disable | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| | amd-gpu-device-plugin | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-109663 addons | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| | disable inspektor-gadget | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-109663 ssh curl -s | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| addons | addons-109663 addons disable | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:35 UTC | 16 Dec 24 10:35 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | addons-109663 addons | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:36 UTC | 16 Dec 24 10:36 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-109663 addons | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:36 UTC | 16 Dec 24 10:36 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-109663 ip | addons-109663 | jenkins | v1.34.0 | 16 Dec 24 10:37 UTC | 16 Dec 24 10:37 UTC |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/12/16 10:32:20
Running on machine: ubuntu-20-agent-4
Binary: Built with gc go1.23.3 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I1216 10:32:20.176960 848599 out.go:345] Setting OutFile to fd 1 ...
I1216 10:32:20.177056 848599 out.go:392] TERM=,COLORTERM=, which probably does not support color
I1216 10:32:20.177064 848599 out.go:358] Setting ErrFile to fd 2...
I1216 10:32:20.177068 848599 out.go:392] TERM=,COLORTERM=, which probably does not support color
I1216 10:32:20.177239 848599 root.go:338] Updating PATH: /home/jenkins/minikube-integration/20107-840384/.minikube/bin
I1216 10:32:20.177825 848599 out.go:352] Setting JSON to false
I1216 10:32:20.178694 848599 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-4","uptime":11687,"bootTime":1734333453,"procs":174,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1071-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I1216 10:32:20.178790 848599 start.go:139] virtualization: kvm guest
I1216 10:32:20.180687 848599 out.go:177] * [addons-109663] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I1216 10:32:20.182103 848599 notify.go:220] Checking for updates...
I1216 10:32:20.182122 848599 out.go:177] - MINIKUBE_LOCATION=20107
I1216 10:32:20.183273 848599 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I1216 10:32:20.184504 848599 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/20107-840384/kubeconfig
I1216 10:32:20.185694 848599 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/20107-840384/.minikube
I1216 10:32:20.186976 848599 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I1216 10:32:20.188067 848599 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I1216 10:32:20.189305 848599 driver.go:394] Setting default libvirt URI to qemu:///system
I1216 10:32:20.210238 848599 docker.go:123] docker version: linux-27.4.0:Docker Engine - Community
I1216 10:32:20.210385 848599 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1216 10:32:20.255369 848599 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-12-16 10:32:20.24671902 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1071-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x8
6_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647935488 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-4 Labels:[] ExperimentalBuild:false ServerVersion:27.4.0 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:88bf19b2105c8b17560993bee28a01ddc2f97182 Expected:88bf19b2105c8b17560993bee28a01ddc2f97182} RuncCommit:{ID:v1.2.2-0-g7cb3632 Expected:v1.2.2-0-g7cb3632} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:[WARNING: bridge
-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled] ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.19.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.31.0] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I1216 10:32:20.255520 848599 docker.go:318] overlay module found
I1216 10:32:20.257178 848599 out.go:177] * Using the docker driver based on user configuration
I1216 10:32:20.258429 848599 start.go:297] selected driver: docker
I1216 10:32:20.258449 848599 start.go:901] validating driver "docker" against <nil>
I1216 10:32:20.258461 848599 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I1216 10:32:20.259277 848599 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I1216 10:32:20.303533 848599 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-12-16 10:32:20.295513369 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1071-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647935488 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-4 Labels:[] ExperimentalBuild:false ServerVersion:27.4.0 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:88bf19b2105c8b17560993bee28a01ddc2f97182 Expected:88bf19b2105c8b17560993bee28a01ddc2f97182} RuncCommit:{ID:v1.2.2-0-g7cb3632 Expected:v1.2.2-0-g7cb3632} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:[WARNING: bridg
e-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled] ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.19.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.31.0] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I1216 10:32:20.303701 848599 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I1216 10:32:20.303936 848599 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I1216 10:32:20.305297 848599 out.go:177] * Using Docker driver with root privileges
I1216 10:32:20.306405 848599 cni.go:84] Creating CNI manager for ""
I1216 10:32:20.306461 848599 cni.go:143] "docker" driver + "crio" runtime found, recommending kindnet
I1216 10:32:20.306471 848599 start_flags.go:319] Found "CNI" CNI - setting NetworkPlugin=cni
I1216 10:32:20.306562 848599 start.go:340] cluster config:
{Name:addons-109663 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.2 ClusterName:addons-109663 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:crio CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.2 ContainerRuntime:crio ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSH
AgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1216 10:32:20.307714 848599 out.go:177] * Starting "addons-109663" primary control-plane node in "addons-109663" cluster
I1216 10:32:20.308731 848599 cache.go:121] Beginning downloading kic base image for docker with crio
I1216 10:32:20.309955 848599 out.go:177] * Pulling base image v0.0.45-1733912881-20083 ...
I1216 10:32:20.311129 848599 preload.go:131] Checking if preload exists for k8s version v1.31.2 and runtime crio
I1216 10:32:20.311157 848599 preload.go:146] Found local preload: /home/jenkins/minikube-integration/20107-840384/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.2-cri-o-overlay-amd64.tar.lz4
I1216 10:32:20.311163 848599 cache.go:56] Caching tarball of preloaded images
I1216 10:32:20.311160 848599 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 in local docker daemon
I1216 10:32:20.311232 848599 preload.go:172] Found /home/jenkins/minikube-integration/20107-840384/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.2-cri-o-overlay-amd64.tar.lz4 in cache, skipping download
I1216 10:32:20.311243 848599 cache.go:59] Finished verifying existence of preloaded tar for v1.31.2 on crio
I1216 10:32:20.311587 848599 profile.go:143] Saving config to /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/config.json ...
I1216 10:32:20.311614 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/config.json: {Name:mkeda270ee12e3e9c2b3f96211254f0d67bf6da1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:20.325703 848599 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 to local cache
I1216 10:32:20.325805 848599 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 in local cache directory
I1216 10:32:20.325824 848599 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 in local cache directory, skipping pull
I1216 10:32:20.325831 848599 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 exists in cache, skipping pull
I1216 10:32:20.325842 848599 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 as a tarball
I1216 10:32:20.325853 848599 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 from local cache
I1216 10:32:32.270421 848599 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 from cached tarball
I1216 10:32:32.270469 848599 cache.go:194] Successfully downloaded all kic artifacts
I1216 10:32:32.270526 848599 start.go:360] acquireMachinesLock for addons-109663: {Name:mk322ac902230420e2cfa3c4d031bb3cb0c61bc8 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I1216 10:32:32.270650 848599 start.go:364] duration metric: took 96.592µs to acquireMachinesLock for "addons-109663"
I1216 10:32:32.270692 848599 start.go:93] Provisioning new machine with config: &{Name:addons-109663 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.2 ClusterName:addons-109663 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:crio CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.2 ContainerRuntime:crio ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQe
muFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.2 ContainerRuntime:crio ControlPlane:true Worker:true}
I1216 10:32:32.270785 848599 start.go:125] createHost starting for "" (driver="docker")
I1216 10:32:32.272482 848599 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I1216 10:32:32.272745 848599 start.go:159] libmachine.API.Create for "addons-109663" (driver="docker")
I1216 10:32:32.272789 848599 client.go:168] LocalClient.Create starting
I1216 10:32:32.272894 848599 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/20107-840384/.minikube/certs/ca.pem
I1216 10:32:32.524572 848599 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/20107-840384/.minikube/certs/cert.pem
I1216 10:32:32.623176 848599 cli_runner.go:164] Run: docker network inspect addons-109663 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W1216 10:32:32.639140 848599 cli_runner.go:211] docker network inspect addons-109663 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I1216 10:32:32.639223 848599 network_create.go:284] running [docker network inspect addons-109663] to gather additional debugging logs...
I1216 10:32:32.639249 848599 cli_runner.go:164] Run: docker network inspect addons-109663
W1216 10:32:32.654825 848599 cli_runner.go:211] docker network inspect addons-109663 returned with exit code 1
I1216 10:32:32.654856 848599 network_create.go:287] error running [docker network inspect addons-109663]: docker network inspect addons-109663: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-109663 not found
I1216 10:32:32.654870 848599 network_create.go:289] output of [docker network inspect addons-109663]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-109663 not found
** /stderr **
I1216 10:32:32.654959 848599 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1216 10:32:32.670365 848599 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc0004f4fa0}
I1216 10:32:32.670414 848599 network_create.go:124] attempt to create docker network addons-109663 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I1216 10:32:32.670452 848599 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-109663 addons-109663
I1216 10:32:32.728359 848599 network_create.go:108] docker network addons-109663 192.168.49.0/24 created
I1216 10:32:32.728388 848599 kic.go:121] calculated static IP "192.168.49.2" for the "addons-109663" container
I1216 10:32:32.728453 848599 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I1216 10:32:32.743748 848599 cli_runner.go:164] Run: docker volume create addons-109663 --label name.minikube.sigs.k8s.io=addons-109663 --label created_by.minikube.sigs.k8s.io=true
I1216 10:32:32.759894 848599 oci.go:103] Successfully created a docker volume addons-109663
I1216 10:32:32.759977 848599 cli_runner.go:164] Run: docker run --rm --name addons-109663-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-109663 --entrypoint /usr/bin/test -v addons-109663:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 -d /var/lib
I1216 10:32:37.657657 848599 cli_runner.go:217] Completed: docker run --rm --name addons-109663-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-109663 --entrypoint /usr/bin/test -v addons-109663:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 -d /var/lib: (4.897631821s)
I1216 10:32:37.657697 848599 oci.go:107] Successfully prepared a docker volume addons-109663
I1216 10:32:37.657718 848599 preload.go:131] Checking if preload exists for k8s version v1.31.2 and runtime crio
I1216 10:32:37.657747 848599 kic.go:194] Starting extracting preloaded images to volume ...
I1216 10:32:37.657821 848599 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/20107-840384/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.2-cri-o-overlay-amd64.tar.lz4:/preloaded.tar:ro -v addons-109663:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 -I lz4 -xf /preloaded.tar -C /extractDir
I1216 10:32:42.147706 848599 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/20107-840384/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.2-cri-o-overlay-amd64.tar.lz4:/preloaded.tar:ro -v addons-109663:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 -I lz4 -xf /preloaded.tar -C /extractDir: (4.489834787s)
I1216 10:32:42.147741 848599 kic.go:203] duration metric: took 4.489992007s to extract preloaded images to volume ...
W1216 10:32:42.147865 848599 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I1216 10:32:42.147983 848599 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I1216 10:32:42.194676 848599 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-109663 --name addons-109663 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-109663 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-109663 --network addons-109663 --ip 192.168.49.2 --volume addons-109663:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2
I1216 10:32:42.492595 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Running}}
I1216 10:32:42.509777 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:32:42.526135 848599 cli_runner.go:164] Run: docker exec addons-109663 stat /var/lib/dpkg/alternatives/iptables
I1216 10:32:42.563632 848599 oci.go:144] the created container "addons-109663" has a running status.
I1216 10:32:42.563664 848599 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa...
I1216 10:32:42.655608 848599 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I1216 10:32:42.674141 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:32:42.690709 848599 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I1216 10:32:42.690729 848599 kic_runner.go:114] Args: [docker exec --privileged addons-109663 chown docker:docker /home/docker/.ssh/authorized_keys]
I1216 10:32:42.733555 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:32:42.751672 848599 machine.go:93] provisionDockerMachine start ...
I1216 10:32:42.751782 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:32:42.769939 848599 main.go:141] libmachine: Using SSH client type: native
I1216 10:32:42.770137 848599 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x867100] 0x869de0 <nil> [] 0s} 127.0.0.1 33139 <nil> <nil>}
I1216 10:32:42.770149 848599 main.go:141] libmachine: About to run SSH command:
hostname
I1216 10:32:42.770885 848599 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: read tcp 127.0.0.1:52954->127.0.0.1:33139: read: connection reset by peer
I1216 10:32:45.894395 848599 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-109663
I1216 10:32:45.894432 848599 ubuntu.go:169] provisioning hostname "addons-109663"
I1216 10:32:45.894492 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:32:45.910952 848599 main.go:141] libmachine: Using SSH client type: native
I1216 10:32:45.911128 848599 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x867100] 0x869de0 <nil> [] 0s} 127.0.0.1 33139 <nil> <nil>}
I1216 10:32:45.911140 848599 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-109663 && echo "addons-109663" | sudo tee /etc/hostname
I1216 10:32:46.045111 848599 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-109663
I1216 10:32:46.045193 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:32:46.061625 848599 main.go:141] libmachine: Using SSH client type: native
I1216 10:32:46.061807 848599 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x867100] 0x869de0 <nil> [] 0s} 127.0.0.1 33139 <nil> <nil>}
I1216 10:32:46.061823 848599 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-109663' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-109663/g' /etc/hosts;
else
echo '127.0.1.1 addons-109663' | sudo tee -a /etc/hosts;
fi
fi
I1216 10:32:46.186853 848599 main.go:141] libmachine: SSH cmd err, output: <nil>:
I1216 10:32:46.186874 848599 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/20107-840384/.minikube CaCertPath:/home/jenkins/minikube-integration/20107-840384/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/20107-840384/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/20107-840384/.minikube}
I1216 10:32:46.186896 848599 ubuntu.go:177] setting up certificates
I1216 10:32:46.186907 848599 provision.go:84] configureAuth start
I1216 10:32:46.186952 848599 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-109663
I1216 10:32:46.201955 848599 provision.go:143] copyHostCerts
I1216 10:32:46.202017 848599 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20107-840384/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/20107-840384/.minikube/ca.pem (1082 bytes)
I1216 10:32:46.202141 848599 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20107-840384/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/20107-840384/.minikube/cert.pem (1123 bytes)
I1216 10:32:46.202206 848599 exec_runner.go:151] cp: /home/jenkins/minikube-integration/20107-840384/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/20107-840384/.minikube/key.pem (1675 bytes)
I1216 10:32:46.202267 848599 provision.go:117] generating server cert: /home/jenkins/minikube-integration/20107-840384/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/20107-840384/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/20107-840384/.minikube/certs/ca-key.pem org=jenkins.addons-109663 san=[127.0.0.1 192.168.49.2 addons-109663 localhost minikube]
I1216 10:32:46.342382 848599 provision.go:177] copyRemoteCerts
I1216 10:32:46.342433 848599 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I1216 10:32:46.342468 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:32:46.358354 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:32:46.448060 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I1216 10:32:46.469314 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I1216 10:32:46.489705 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I1216 10:32:46.509732 848599 provision.go:87] duration metric: took 322.814241ms to configureAuth
I1216 10:32:46.509759 848599 ubuntu.go:193] setting minikube options for container-runtime
I1216 10:32:46.509910 848599 config.go:182] Loaded profile config "addons-109663": Driver=docker, ContainerRuntime=crio, KubernetesVersion=v1.31.2
I1216 10:32:46.510000 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:32:46.526470 848599 main.go:141] libmachine: Using SSH client type: native
I1216 10:32:46.526646 848599 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x867100] 0x869de0 <nil> [] 0s} 127.0.0.1 33139 <nil> <nil>}
I1216 10:32:46.526667 848599 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /etc/sysconfig && printf %s "
CRIO_MINIKUBE_OPTIONS='--insecure-registry 10.96.0.0/12 '
" | sudo tee /etc/sysconfig/crio.minikube && sudo systemctl restart crio
I1216 10:32:46.731259 848599 main.go:141] libmachine: SSH cmd err, output: <nil>:
CRIO_MINIKUBE_OPTIONS='--insecure-registry 10.96.0.0/12 '
I1216 10:32:46.731288 848599 machine.go:96] duration metric: took 3.97958665s to provisionDockerMachine
I1216 10:32:46.731304 848599 client.go:171] duration metric: took 14.458503354s to LocalClient.Create
I1216 10:32:46.731327 848599 start.go:167] duration metric: took 14.458580941s to libmachine.API.Create "addons-109663"
I1216 10:32:46.731337 848599 start.go:293] postStartSetup for "addons-109663" (driver="docker")
I1216 10:32:46.731348 848599 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I1216 10:32:46.731400 848599 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I1216 10:32:46.731446 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:32:46.748035 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:32:46.835559 848599 ssh_runner.go:195] Run: cat /etc/os-release
I1216 10:32:46.838341 848599 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I1216 10:32:46.838368 848599 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I1216 10:32:46.838385 848599 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I1216 10:32:46.838394 848599 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I1216 10:32:46.838411 848599 filesync.go:126] Scanning /home/jenkins/minikube-integration/20107-840384/.minikube/addons for local assets ...
I1216 10:32:46.838464 848599 filesync.go:126] Scanning /home/jenkins/minikube-integration/20107-840384/.minikube/files for local assets ...
I1216 10:32:46.838507 848599 start.go:296] duration metric: took 107.161933ms for postStartSetup
I1216 10:32:46.838809 848599 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-109663
I1216 10:32:46.854233 848599 profile.go:143] Saving config to /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/config.json ...
I1216 10:32:46.854469 848599 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I1216 10:32:46.854512 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:32:46.869838 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:32:46.959575 848599 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I1216 10:32:46.963386 848599 start.go:128] duration metric: took 14.692586018s to createHost
I1216 10:32:46.963416 848599 start.go:83] releasing machines lock for "addons-109663", held for 14.692749507s
I1216 10:32:46.963496 848599 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-109663
I1216 10:32:46.978640 848599 ssh_runner.go:195] Run: cat /version.json
I1216 10:32:46.978677 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:32:46.978701 848599 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I1216 10:32:46.978764 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:32:46.995635 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:32:46.996158 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:32:47.078650 848599 ssh_runner.go:195] Run: systemctl --version
I1216 10:32:47.143743 848599 ssh_runner.go:195] Run: sudo sh -c "podman version >/dev/null"
I1216 10:32:47.278506 848599 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I1216 10:32:47.282415 848599 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "sudo mv {} {}.mk_disabled" ;
I1216 10:32:47.299071 848599 cni.go:221] loopback cni configuration disabled: "/etc/cni/net.d/*loopback.conf*" found
I1216 10:32:47.299151 848599 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I1216 10:32:47.324690 848599 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I1216 10:32:47.324713 848599 start.go:495] detecting cgroup driver to use...
I1216 10:32:47.324748 848599 detect.go:187] detected "cgroupfs" cgroup driver on host os
I1216 10:32:47.324785 848599 ssh_runner.go:195] Run: sudo systemctl stop -f containerd
I1216 10:32:47.338062 848599 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service containerd
I1216 10:32:47.347179 848599 docker.go:217] disabling cri-docker service (if available) ...
I1216 10:32:47.347216 848599 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
I1216 10:32:47.358730 848599 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
I1216 10:32:47.370504 848599 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
I1216 10:32:47.449813 848599 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
I1216 10:32:47.518777 848599 docker.go:233] disabling docker service ...
I1216 10:32:47.518823 848599 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
I1216 10:32:47.535753 848599 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
I1216 10:32:47.545023 848599 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
I1216 10:32:47.617016 848599 ssh_runner.go:195] Run: sudo systemctl mask docker.service
I1216 10:32:47.691960 848599 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I1216 10:32:47.701141 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/crio/crio.sock
" | sudo tee /etc/crictl.yaml"
I1216 10:32:47.714441 848599 crio.go:59] configure cri-o to use "registry.k8s.io/pause:3.10" pause image...
I1216 10:32:47.714485 848599 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|^.*pause_image = .*$|pause_image = "registry.k8s.io/pause:3.10"|' /etc/crio/crio.conf.d/02-crio.conf"
I1216 10:32:47.722745 848599 crio.go:70] configuring cri-o to use "cgroupfs" as cgroup driver...
I1216 10:32:47.722785 848599 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|^.*cgroup_manager = .*$|cgroup_manager = "cgroupfs"|' /etc/crio/crio.conf.d/02-crio.conf"
I1216 10:32:47.731282 848599 ssh_runner.go:195] Run: sh -c "sudo sed -i '/conmon_cgroup = .*/d' /etc/crio/crio.conf.d/02-crio.conf"
I1216 10:32:47.739260 848599 ssh_runner.go:195] Run: sh -c "sudo sed -i '/cgroup_manager = .*/a conmon_cgroup = "pod"' /etc/crio/crio.conf.d/02-crio.conf"
I1216 10:32:47.747136 848599 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I1216 10:32:47.754606 848599 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *"net.ipv4.ip_unprivileged_port_start=.*"/d' /etc/crio/crio.conf.d/02-crio.conf"
I1216 10:32:47.762450 848599 ssh_runner.go:195] Run: sh -c "sudo grep -q "^ *default_sysctls" /etc/crio/crio.conf.d/02-crio.conf || sudo sed -i '/conmon_cgroup = .*/a default_sysctls = \[\n\]' /etc/crio/crio.conf.d/02-crio.conf"
I1216 10:32:47.775303 848599 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^default_sysctls *= *\[|&\n "net.ipv4.ip_unprivileged_port_start=0",|' /etc/crio/crio.conf.d/02-crio.conf"
I1216 10:32:47.783168 848599 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I1216 10:32:47.789866 848599 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I1216 10:32:47.796667 848599 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1216 10:32:47.867618 848599 ssh_runner.go:195] Run: sudo systemctl restart crio
I1216 10:32:47.965258 848599 start.go:542] Will wait 60s for socket path /var/run/crio/crio.sock
I1216 10:32:47.965321 848599 ssh_runner.go:195] Run: stat /var/run/crio/crio.sock
I1216 10:32:47.968509 848599 start.go:563] Will wait 60s for crictl version
I1216 10:32:47.968560 848599 ssh_runner.go:195] Run: which crictl
I1216 10:32:47.971345 848599 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I1216 10:32:48.003861 848599 start.go:579] Version: 0.1.0
RuntimeName: cri-o
RuntimeVersion: 1.24.6
RuntimeApiVersion: v1
I1216 10:32:48.003945 848599 ssh_runner.go:195] Run: crio --version
I1216 10:32:48.037600 848599 ssh_runner.go:195] Run: crio --version
I1216 10:32:48.070624 848599 out.go:177] * Preparing Kubernetes v1.31.2 on CRI-O 1.24.6 ...
I1216 10:32:48.071731 848599 cli_runner.go:164] Run: docker network inspect addons-109663 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I1216 10:32:48.086790 848599 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I1216 10:32:48.089949 848599 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I1216 10:32:48.099631 848599 kubeadm.go:883] updating cluster {Name:addons-109663 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.2 ClusterName:addons-109663 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:crio CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.2 ContainerRuntime:crio ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmw
arePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I1216 10:32:48.099753 848599 preload.go:131] Checking if preload exists for k8s version v1.31.2 and runtime crio
I1216 10:32:48.099811 848599 ssh_runner.go:195] Run: sudo crictl images --output json
I1216 10:32:48.162586 848599 crio.go:514] all images are preloaded for cri-o runtime.
I1216 10:32:48.162609 848599 crio.go:433] Images already preloaded, skipping extraction
I1216 10:32:48.162661 848599 ssh_runner.go:195] Run: sudo crictl images --output json
I1216 10:32:48.193727 848599 crio.go:514] all images are preloaded for cri-o runtime.
I1216 10:32:48.193748 848599 cache_images.go:84] Images are preloaded, skipping loading
I1216 10:32:48.193759 848599 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.2 crio true true} ...
I1216 10:32:48.193856 848599 kubeadm.go:946] kubelet [Unit]
Wants=crio.service
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.2/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --cgroups-per-qos=false --config=/var/lib/kubelet/config.yaml --enforce-node-allocatable= --hostname-override=addons-109663 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.2 ClusterName:addons-109663 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:crio CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I1216 10:32:48.193930 848599 ssh_runner.go:195] Run: crio config
I1216 10:32:48.233450 848599 cni.go:84] Creating CNI manager for ""
I1216 10:32:48.233469 848599 cni.go:143] "docker" driver + "crio" runtime found, recommending kindnet
I1216 10:32:48.233479 848599 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I1216 10:32:48.233499 848599 kubeadm.go:189] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.2 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-109663 NodeName:addons-109663 DNSDomain:cluster.local CRISocket:/var/run/crio/crio.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuberne
tes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/crio/crio.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I1216 10:32:48.233626 848599 kubeadm.go:195] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta4
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/crio/crio.sock
name: "addons-109663"
kubeletExtraArgs:
- name: "node-ip"
value: "192.168.49.2"
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta4
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
- name: "enable-admission-plugins"
value: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
- name: "allocate-node-cidrs"
value: "true"
- name: "leader-elect"
value: "false"
scheduler:
extraArgs:
- name: "leader-elect"
value: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
- name: "proxy-refresh-interval"
value: "70000"
kubernetesVersion: v1.31.2
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/crio/crio.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I1216 10:32:48.233678 848599 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.2
I1216 10:32:48.241287 848599 binaries.go:44] Found k8s binaries, skipping transfer
I1216 10:32:48.241353 848599 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I1216 10:32:48.248690 848599 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (363 bytes)
I1216 10:32:48.263749 848599 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I1216 10:32:48.278768 848599 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2287 bytes)
I1216 10:32:48.293446 848599 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I1216 10:32:48.296360 848599 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I1216 10:32:48.305456 848599 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1216 10:32:48.385855 848599 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1216 10:32:48.396926 848599 certs.go:68] Setting up /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663 for IP: 192.168.49.2
I1216 10:32:48.396946 848599 certs.go:194] generating shared ca certs ...
I1216 10:32:48.396972 848599 certs.go:226] acquiring lock for ca certs: {Name:mkc11fd68d423e1cca90bec28435e0a6c7ecf1c1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:48.397158 848599 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/20107-840384/.minikube/ca.key
I1216 10:32:48.466160 848599 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20107-840384/.minikube/ca.crt ...
I1216 10:32:48.466182 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/ca.crt: {Name:mk1859f6bdff9985876c6f50db5f2d1280c287c0 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:48.466320 848599 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20107-840384/.minikube/ca.key ...
I1216 10:32:48.466340 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/ca.key: {Name:mk92e534493378752c6e08cd41ae73570fe64ae3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:48.466434 848599 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/20107-840384/.minikube/proxy-client-ca.key
I1216 10:32:48.531063 848599 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20107-840384/.minikube/proxy-client-ca.crt ...
I1216 10:32:48.531083 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/proxy-client-ca.crt: {Name:mkbd2c16dce66b8bd8800e09edb15d99e74a3dee Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:48.531214 848599 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20107-840384/.minikube/proxy-client-ca.key ...
I1216 10:32:48.531227 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/proxy-client-ca.key: {Name:mkbd79e00e7ff3c72871d6c44df9bbc55c8438ac Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:48.531295 848599 certs.go:256] generating profile certs ...
I1216 10:32:48.531350 848599 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/client.key
I1216 10:32:48.531370 848599 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/client.crt with IP's: []
I1216 10:32:48.582934 848599 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/client.crt ...
I1216 10:32:48.582951 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/client.crt: {Name:mk7800253813d63a2b9feff6a9f93fbd096ed71c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:48.583055 848599 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/client.key ...
I1216 10:32:48.583065 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/client.key: {Name:mk68a33bf12fa88f4decce469c4693c84cfbbe9b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:48.583137 848599 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.key.5a4e409c
I1216 10:32:48.583153 848599 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.crt.5a4e409c with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I1216 10:32:48.795073 848599 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.crt.5a4e409c ...
I1216 10:32:48.795097 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.crt.5a4e409c: {Name:mkb1a96aec38a507038981d80b8c62dd0085ece6 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:48.795228 848599 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.key.5a4e409c ...
I1216 10:32:48.795240 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.key.5a4e409c: {Name:mk0eaba652e54fc0326310c214d334efd837fdd7 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:48.795306 848599 certs.go:381] copying /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.crt.5a4e409c -> /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.crt
I1216 10:32:48.795380 848599 certs.go:385] copying /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.key.5a4e409c -> /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.key
I1216 10:32:48.795425 848599 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/proxy-client.key
I1216 10:32:48.795441 848599 crypto.go:68] Generating cert /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/proxy-client.crt with IP's: []
I1216 10:32:49.185378 848599 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/proxy-client.crt ...
I1216 10:32:49.185403 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/proxy-client.crt: {Name:mkfb8512dec95af5f7fe9be594be404ecbc3feb4 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:49.185541 848599 crypto.go:164] Writing key to /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/proxy-client.key ...
I1216 10:32:49.185553 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/proxy-client.key: {Name:mk1e40828877323680e6bc49b0f353b0f4a8d014 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:32:49.185723 848599 certs.go:484] found cert: /home/jenkins/minikube-integration/20107-840384/.minikube/certs/ca-key.pem (1679 bytes)
I1216 10:32:49.185757 848599 certs.go:484] found cert: /home/jenkins/minikube-integration/20107-840384/.minikube/certs/ca.pem (1082 bytes)
I1216 10:32:49.185782 848599 certs.go:484] found cert: /home/jenkins/minikube-integration/20107-840384/.minikube/certs/cert.pem (1123 bytes)
I1216 10:32:49.185813 848599 certs.go:484] found cert: /home/jenkins/minikube-integration/20107-840384/.minikube/certs/key.pem (1675 bytes)
I1216 10:32:49.186471 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I1216 10:32:49.208462 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I1216 10:32:49.229862 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I1216 10:32:49.252897 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I1216 10:32:49.272707 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I1216 10:32:49.292707 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I1216 10:32:49.312720 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I1216 10:32:49.332990 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/profiles/addons-109663/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I1216 10:32:49.352947 848599 ssh_runner.go:362] scp /home/jenkins/minikube-integration/20107-840384/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I1216 10:32:49.372714 848599 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I1216 10:32:49.387254 848599 ssh_runner.go:195] Run: openssl version
I1216 10:32:49.391984 848599 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I1216 10:32:49.400361 848599 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I1216 10:32:49.403124 848599 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Dec 16 10:32 /usr/share/ca-certificates/minikubeCA.pem
I1216 10:32:49.403185 848599 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I1216 10:32:49.409075 848599 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I1216 10:32:49.416721 848599 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I1216 10:32:49.419513 848599 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I1216 10:32:49.419554 848599 kubeadm.go:392] StartCluster: {Name:addons-109663 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1733912881-20083@sha256:64d8b27f78fd269d886e21ba8fc88be20de183ba5cc5bce33d0810e8a65f1df2 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.2 ClusterName:addons-109663 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:crio CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.2 ContainerRuntime:crio ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmware
Path: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I1216 10:32:49.419649 848599 cri.go:54] listing CRI containers in root : {State:paused Name: Namespaces:[kube-system]}
I1216 10:32:49.419715 848599 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
I1216 10:32:49.451626 848599 cri.go:89] found id: ""
I1216 10:32:49.451684 848599 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I1216 10:32:49.459035 848599 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I1216 10:32:49.466323 848599 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I1216 10:32:49.466368 848599 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I1216 10:32:49.473764 848599 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I1216 10:32:49.473786 848599 kubeadm.go:157] found existing configuration files:
I1216 10:32:49.473827 848599 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I1216 10:32:49.481337 848599 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I1216 10:32:49.481398 848599 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I1216 10:32:49.488336 848599 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I1216 10:32:49.495564 848599 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I1216 10:32:49.495613 848599 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I1216 10:32:49.502569 848599 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I1216 10:32:49.509725 848599 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I1216 10:32:49.509771 848599 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I1216 10:32:49.517196 848599 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I1216 10:32:49.524832 848599 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I1216 10:32:49.524874 848599 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I1216 10:32:49.531758 848599 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.2:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I1216 10:32:49.565735 848599 kubeadm.go:310] [init] Using Kubernetes version: v1.31.2
I1216 10:32:49.565811 848599 kubeadm.go:310] [preflight] Running pre-flight checks
I1216 10:32:49.580542 848599 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I1216 10:32:49.580609 848599 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1071-gcp[0m
I1216 10:32:49.580646 848599 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I1216 10:32:49.580689 848599 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I1216 10:32:49.580778 848599 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I1216 10:32:49.580831 848599 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I1216 10:32:49.580871 848599 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I1216 10:32:49.580933 848599 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I1216 10:32:49.581006 848599 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I1216 10:32:49.581096 848599 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I1216 10:32:49.581174 848599 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I1216 10:32:49.581246 848599 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I1216 10:32:49.629850 848599 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I1216 10:32:49.630004 848599 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I1216 10:32:49.630169 848599 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I1216 10:32:49.636020 848599 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I1216 10:32:49.639441 848599 out.go:235] - Generating certificates and keys ...
I1216 10:32:49.639556 848599 kubeadm.go:310] [certs] Using existing ca certificate authority
I1216 10:32:49.639619 848599 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I1216 10:32:49.836390 848599 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I1216 10:32:50.046656 848599 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I1216 10:32:50.397346 848599 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I1216 10:32:50.460502 848599 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I1216 10:32:50.635424 848599 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I1216 10:32:50.635586 848599 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-109663 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I1216 10:32:50.820560 848599 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I1216 10:32:50.820691 848599 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-109663 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I1216 10:32:51.004936 848599 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I1216 10:32:51.062758 848599 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I1216 10:32:51.170931 848599 kubeadm.go:310] [certs] Generating "sa" key and public key
I1216 10:32:51.170996 848599 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I1216 10:32:51.335077 848599 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I1216 10:32:51.557386 848599 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I1216 10:32:51.984782 848599 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I1216 10:32:52.326144 848599 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I1216 10:32:52.700266 848599 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I1216 10:32:52.700739 848599 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I1216 10:32:52.703004 848599 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I1216 10:32:52.704967 848599 out.go:235] - Booting up control plane ...
I1216 10:32:52.705051 848599 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I1216 10:32:52.705133 848599 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I1216 10:32:52.705698 848599 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I1216 10:32:52.714141 848599 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I1216 10:32:52.718979 848599 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I1216 10:32:52.719037 848599 kubeadm.go:310] [kubelet-start] Starting the kubelet
I1216 10:32:52.797628 848599 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I1216 10:32:52.797772 848599 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I1216 10:32:53.298364 848599 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 500.804305ms
I1216 10:32:53.298473 848599 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I1216 10:32:57.300009 848599 kubeadm.go:310] [api-check] The API server is healthy after 4.001662107s
I1216 10:32:57.310037 848599 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I1216 10:32:57.319228 848599 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I1216 10:32:57.334100 848599 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I1216 10:32:57.334362 848599 kubeadm.go:310] [mark-control-plane] Marking the node addons-109663 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I1216 10:32:57.340771 848599 kubeadm.go:310] [bootstrap-token] Using token: 2h4i74.yidhy7fpg06tydg2
I1216 10:32:57.341964 848599 out.go:235] - Configuring RBAC rules ...
I1216 10:32:57.342133 848599 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I1216 10:32:57.345049 848599 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I1216 10:32:57.350019 848599 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I1216 10:32:57.352300 848599 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I1216 10:32:57.355264 848599 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I1216 10:32:57.357395 848599 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I1216 10:32:57.705816 848599 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I1216 10:32:58.121923 848599 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I1216 10:32:58.704608 848599 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I1216 10:32:58.705594 848599 kubeadm.go:310]
I1216 10:32:58.705701 848599 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I1216 10:32:58.705720 848599 kubeadm.go:310]
I1216 10:32:58.705826 848599 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I1216 10:32:58.705837 848599 kubeadm.go:310]
I1216 10:32:58.705873 848599 kubeadm.go:310] mkdir -p $HOME/.kube
I1216 10:32:58.705959 848599 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I1216 10:32:58.706029 848599 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I1216 10:32:58.706038 848599 kubeadm.go:310]
I1216 10:32:58.706098 848599 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I1216 10:32:58.706107 848599 kubeadm.go:310]
I1216 10:32:58.706168 848599 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I1216 10:32:58.706183 848599 kubeadm.go:310]
I1216 10:32:58.706227 848599 kubeadm.go:310] You should now deploy a pod network to the cluster.
I1216 10:32:58.706298 848599 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I1216 10:32:58.706360 848599 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I1216 10:32:58.706369 848599 kubeadm.go:310]
I1216 10:32:58.706437 848599 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I1216 10:32:58.706507 848599 kubeadm.go:310] and service account keys on each node and then running the following as root:
I1216 10:32:58.706514 848599 kubeadm.go:310]
I1216 10:32:58.706586 848599 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 2h4i74.yidhy7fpg06tydg2 \
I1216 10:32:58.706682 848599 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:e6a6471f4384e10723e2292fb8d114ab4ea25aa738d7f29c5187bb98e939b6b4 \
I1216 10:32:58.706706 848599 kubeadm.go:310] --control-plane
I1216 10:32:58.706718 848599 kubeadm.go:310]
I1216 10:32:58.706818 848599 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I1216 10:32:58.706831 848599 kubeadm.go:310]
I1216 10:32:58.706927 848599 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 2h4i74.yidhy7fpg06tydg2 \
I1216 10:32:58.707051 848599 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:e6a6471f4384e10723e2292fb8d114ab4ea25aa738d7f29c5187bb98e939b6b4
I1216 10:32:58.709379 848599 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1071-gcp\n", err: exit status 1
I1216 10:32:58.709495 848599 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I1216 10:32:58.709516 848599 cni.go:84] Creating CNI manager for ""
I1216 10:32:58.709524 848599 cni.go:143] "docker" driver + "crio" runtime found, recommending kindnet
I1216 10:32:58.711000 848599 out.go:177] * Configuring CNI (Container Networking Interface) ...
I1216 10:32:58.712127 848599 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
I1216 10:32:58.715765 848599 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.31.2/kubectl ...
I1216 10:32:58.715784 848599 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2601 bytes)
I1216 10:32:58.731953 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
I1216 10:32:58.917087 848599 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I1216 10:32:58.917200 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:32:58.917234 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-109663 minikube.k8s.io/updated_at=2024_12_16T10_32_58_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=22da80be3b90f71512d84256b3df4ef76bd13ff8 minikube.k8s.io/name=addons-109663 minikube.k8s.io/primary=true
I1216 10:32:58.924437 848599 ops.go:34] apiserver oom_adj: -16
I1216 10:32:58.985975 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:32:59.486347 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:32:59.986971 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:33:00.486306 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:33:00.986656 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:33:01.486381 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:33:01.986792 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:33:02.486214 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:33:02.986029 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:33:03.486249 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:33:03.986520 848599 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I1216 10:33:04.087749 848599 kubeadm.go:1113] duration metric: took 5.170647368s to wait for elevateKubeSystemPrivileges
I1216 10:33:04.087800 848599 kubeadm.go:394] duration metric: took 14.668249445s to StartCluster
I1216 10:33:04.087826 848599 settings.go:142] acquiring lock: {Name:mk06b7df26b8c35e37c6f668a6089af3b5005238 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:33:04.087950 848599 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/20107-840384/kubeconfig
I1216 10:33:04.088601 848599 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/20107-840384/kubeconfig: {Name:mkf0f71705623f4096af1601d96997d88188e951 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I1216 10:33:04.088814 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I1216 10:33:04.088833 848599 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.2 ContainerRuntime:crio ControlPlane:true Worker:true}
I1216 10:33:04.088909 848599 addons.go:507] enable addons start: toEnable=map[ambassador:false amd-gpu-device-plugin:true auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I1216 10:33:04.089044 848599 addons.go:69] Setting yakd=true in profile "addons-109663"
I1216 10:33:04.089083 848599 addons.go:234] Setting addon yakd=true in "addons-109663"
I1216 10:33:04.089098 848599 addons.go:69] Setting inspektor-gadget=true in profile "addons-109663"
I1216 10:33:04.089104 848599 config.go:182] Loaded profile config "addons-109663": Driver=docker, ContainerRuntime=crio, KubernetesVersion=v1.31.2
I1216 10:33:04.089121 848599 addons.go:234] Setting addon inspektor-gadget=true in "addons-109663"
I1216 10:33:04.089133 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.089128 848599 addons.go:69] Setting default-storageclass=true in profile "addons-109663"
I1216 10:33:04.089145 848599 addons.go:69] Setting cloud-spanner=true in profile "addons-109663"
I1216 10:33:04.089171 848599 addons.go:69] Setting amd-gpu-device-plugin=true in profile "addons-109663"
I1216 10:33:04.089173 848599 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-109663"
I1216 10:33:04.089187 848599 addons.go:234] Setting addon amd-gpu-device-plugin=true in "addons-109663"
I1216 10:33:04.089194 848599 addons.go:69] Setting ingress=true in profile "addons-109663"
I1216 10:33:04.089197 848599 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-109663"
I1216 10:33:04.089213 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.089219 848599 addons.go:234] Setting addon ingress=true in "addons-109663"
I1216 10:33:04.089236 848599 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-109663"
I1216 10:33:04.089256 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.089270 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.089587 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.089704 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.089738 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.089751 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.089756 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.089986 848599 addons.go:69] Setting ingress-dns=true in profile "addons-109663"
I1216 10:33:04.090008 848599 addons.go:234] Setting addon ingress-dns=true in "addons-109663"
I1216 10:33:04.090019 848599 addons.go:69] Setting storage-provisioner=true in profile "addons-109663"
I1216 10:33:04.090042 848599 addons.go:234] Setting addon storage-provisioner=true in "addons-109663"
I1216 10:33:04.090056 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.090072 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.090107 848599 addons.go:69] Setting volcano=true in profile "addons-109663"
I1216 10:33:04.090146 848599 addons.go:234] Setting addon volcano=true in "addons-109663"
I1216 10:33:04.090170 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.090589 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.090631 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.090645 848599 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-109663"
I1216 10:33:04.090662 848599 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-109663"
I1216 10:33:04.090912 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.091232 848599 out.go:177] * Verifying Kubernetes components...
I1216 10:33:04.089187 848599 addons.go:234] Setting addon cloud-spanner=true in "addons-109663"
I1216 10:33:04.091364 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.089180 848599 addons.go:69] Setting gcp-auth=true in profile "addons-109663"
I1216 10:33:04.091483 848599 mustload.go:65] Loading cluster: addons-109663
I1216 10:33:04.091537 848599 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-109663"
I1216 10:33:04.091591 848599 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-109663"
I1216 10:33:04.091635 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.091761 848599 config.go:182] Loaded profile config "addons-109663": Driver=docker, ContainerRuntime=crio, KubernetesVersion=v1.31.2
I1216 10:33:04.091927 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.092030 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.092131 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.095721 848599 addons.go:69] Setting metrics-server=true in profile "addons-109663"
I1216 10:33:04.095747 848599 addons.go:234] Setting addon metrics-server=true in "addons-109663"
I1216 10:33:04.095746 848599 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I1216 10:33:04.095777 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.095868 848599 addons.go:69] Setting registry=true in profile "addons-109663"
I1216 10:33:04.091362 848599 addons.go:69] Setting volumesnapshots=true in profile "addons-109663"
I1216 10:33:04.089169 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.095956 848599 addons.go:234] Setting addon registry=true in "addons-109663"
I1216 10:33:04.095996 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.096414 848599 addons.go:234] Setting addon volumesnapshots=true in "addons-109663"
I1216 10:33:04.096445 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.096452 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.096755 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.096949 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.097026 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.090631 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.130675 848599 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I1216 10:33:04.132180 848599 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I1216 10:33:04.132204 848599 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I1216 10:33:04.132270 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.146511 848599 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.8
I1216 10:33:04.146594 848599 out.go:177] - Using image docker.io/rocm/k8s-device-plugin:1.25.2.8
I1216 10:33:04.147723 848599 addons.go:431] installing /etc/kubernetes/addons/amd-gpu-device-plugin.yaml
I1216 10:33:04.147747 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/amd-gpu-device-plugin.yaml (1868 bytes)
I1216 10:33:04.147825 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.149145 848599 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I1216 10:33:04.150251 848599 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I1216 10:33:04.150478 848599 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-109663"
I1216 10:33:04.150545 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.150630 848599 out.go:177] - Using image docker.io/registry:2.8.3
I1216 10:33:04.151011 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.152794 848599 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I1216 10:33:04.152887 848599 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.25
I1216 10:33:04.154036 848599 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I1216 10:33:04.154055 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I1216 10:33:04.154111 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.154856 848599 addons.go:234] Setting addon default-storageclass=true in "addons-109663"
I1216 10:33:04.154903 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.155351 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:04.155923 848599 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I1216 10:33:04.156321 848599 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I1216 10:33:04.156342 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I1216 10:33:04.156387 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.158101 848599 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I1216 10:33:04.159206 848599 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.4
I1216 10:33:04.160412 848599 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I1216 10:33:04.161739 848599 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.4
I1216 10:33:04.162741 848599 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I1216 10:33:04.163984 848599 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.3
I1216 10:33:04.164963 848599 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I1216 10:33:04.165572 848599 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I1216 10:33:04.165607 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I1216 10:33:04.165667 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.167879 848599 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I1216 10:33:04.170258 848599 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I1216 10:33:04.170302 848599 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
W1216 10:33:04.170310 848599 out.go:270] ! Enabling 'volcano' returned an error: running callbacks: [volcano addon does not support crio]
I1216 10:33:04.170277 848599 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.35.0
I1216 10:33:04.170471 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.172519 848599 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I1216 10:33:04.172537 848599 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5248 bytes)
I1216 10:33:04.172605 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.175569 848599 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I1216 10:33:04.175587 848599 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I1216 10:33:04.175660 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.210930 848599 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I1216 10:33:04.212499 848599 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I1216 10:33:04.212530 848599 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I1216 10:33:04.212615 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.235651 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.235699 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.236358 848599 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I1216 10:33:04.236380 848599 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I1216 10:33:04.236442 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.236549 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.236651 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:04.240224 848599 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I1216 10:33:04.240241 848599 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I1216 10:33:04.240414 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.240883 848599 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.17.0
I1216 10:33:04.241768 848599 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I1216 10:33:04.241792 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I1216 10:33:04.241845 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.242344 848599 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I1216 10:33:04.242364 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I1216 10:33:04.242425 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.243294 848599 out.go:177] - Using image docker.io/busybox:stable
I1216 10:33:04.244353 848599 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I1216 10:33:04.244382 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I1216 10:33:04.244429 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.244885 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.245310 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.246878 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.246930 848599 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I1216 10:33:04.248030 848599 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I1216 10:33:04.248051 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I1216 10:33:04.248105 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:04.263238 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.271380 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.271660 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.272475 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
W1216 10:33:04.279597 848599 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I1216 10:33:04.279627 848599 retry.go:31] will retry after 144.207623ms: ssh: handshake failed: EOF
I1216 10:33:04.288206 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.296070 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I1216 10:33:04.296458 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:04.296707 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
W1216 10:33:04.297011 848599 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I1216 10:33:04.297031 848599 retry.go:31] will retry after 279.355591ms: ssh: handshake failed: EOF
I1216 10:33:04.494587 848599 ssh_runner.go:195] Run: sudo systemctl start kubelet
I1216 10:33:04.589727 848599 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I1216 10:33:04.589761 848599 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I1216 10:33:04.673229 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I1216 10:33:04.679498 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I1216 10:33:04.774125 848599 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I1216 10:33:04.774165 848599 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I1216 10:33:04.777542 848599 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I1216 10:33:04.777581 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I1216 10:33:04.783458 848599 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I1216 10:33:04.783503 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I1216 10:33:04.794089 848599 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I1216 10:33:04.794123 848599 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I1216 10:33:04.794663 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/amd-gpu-device-plugin.yaml
I1216 10:33:04.873178 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I1216 10:33:04.879090 848599 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I1216 10:33:04.879113 848599 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I1216 10:33:04.881997 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I1216 10:33:04.885458 848599 addons.go:431] installing /etc/kubernetes/addons/ig-deployment.yaml
I1216 10:33:04.885483 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-deployment.yaml (14576 bytes)
I1216 10:33:04.892271 848599 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I1216 10:33:04.892293 848599 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I1216 10:33:04.895905 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I1216 10:33:04.972830 848599 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I1216 10:33:04.972868 848599 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I1216 10:33:04.973550 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I1216 10:33:04.979372 848599 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I1216 10:33:04.979401 848599 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I1216 10:33:05.074625 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I1216 10:33:05.076930 848599 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I1216 10:33:05.076956 848599 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I1216 10:33:05.186319 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I1216 10:33:05.188659 848599 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I1216 10:33:05.188699 848599 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I1216 10:33:05.191024 848599 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I1216 10:33:05.191062 848599 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I1216 10:33:05.289209 848599 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I1216 10:33:05.289291 848599 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I1216 10:33:05.290410 848599 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I1216 10:33:05.290476 848599 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I1216 10:33:05.372799 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-deployment.yaml
I1216 10:33:05.384174 848599 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I1216 10:33:05.384201 848599 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I1216 10:33:05.592500 848599 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I1216 10:33:05.592549 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I1216 10:33:05.773126 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I1216 10:33:05.879400 848599 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I1216 10:33:05.879511 848599 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I1216 10:33:05.888426 848599 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I1216 10:33:05.888459 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I1216 10:33:06.077724 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I1216 10:33:06.178196 848599 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.882086183s)
I1216 10:33:06.178247 848599 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I1216 10:33:06.179598 848599 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.684975957s)
I1216 10:33:06.180502 848599 node_ready.go:35] waiting up to 6m0s for node "addons-109663" to be "Ready" ...
I1216 10:33:06.373531 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I1216 10:33:06.389254 848599 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I1216 10:33:06.389294 848599 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I1216 10:33:06.898854 848599 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-109663" context rescaled to 1 replicas
I1216 10:33:06.985246 848599 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I1216 10:33:06.985287 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I1216 10:33:07.274010 848599 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I1216 10:33:07.274100 848599 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I1216 10:33:07.472478 848599 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I1216 10:33:07.472515 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I1216 10:33:07.687775 848599 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I1216 10:33:07.687864 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I1216 10:33:07.974316 848599 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I1216 10:33:07.974399 848599 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I1216 10:33:08.274679 848599 node_ready.go:53] node "addons-109663" has status "Ready":"False"
I1216 10:33:08.287244 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I1216 10:33:09.273222 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (4.599944933s)
I1216 10:33:10.480914 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (5.801378523s)
I1216 10:33:10.480957 848599 addons.go:475] Verifying addon ingress=true in "addons-109663"
I1216 10:33:10.481001 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/amd-gpu-device-plugin.yaml: (5.686304837s)
I1216 10:33:10.481109 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (5.607900487s)
I1216 10:33:10.481203 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (5.599182204s)
I1216 10:33:10.481438 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (5.585508598s)
I1216 10:33:10.481459 848599 addons.go:475] Verifying addon registry=true in "addons-109663"
I1216 10:33:10.481883 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-deployment.yaml: (5.109009356s)
I1216 10:33:10.481668 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (5.508095231s)
I1216 10:33:10.481710 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (5.407059386s)
I1216 10:33:10.481778 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (5.295416881s)
I1216 10:33:10.481971 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (4.708752789s)
I1216 10:33:10.481991 848599 addons.go:475] Verifying addon metrics-server=true in "addons-109663"
I1216 10:33:10.482371 848599 out.go:177] * Verifying ingress addon...
I1216 10:33:10.483212 848599 out.go:177] * Verifying registry addon...
I1216 10:33:10.484880 848599 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I1216 10:33:10.485888 848599 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I1216 10:33:10.491352 848599 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I1216 10:33:10.491375 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:10.492370 848599 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=registry
I1216 10:33:10.492394 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
W1216 10:33:10.497556 848599 out.go:270] ! Enabling 'storage-provisioner-rancher' returned an error: running callbacks: [Error making local-path the default storage class: Error while marking storage class local-path as default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
I1216 10:33:10.685455 848599 node_ready.go:53] node "addons-109663" has status "Ready":"False"
I1216 10:33:10.989291 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:10.990123 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:11.410531 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (5.332755094s)
W1216 10:33:11.410572 848599 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I1216 10:33:11.410594 848599 retry.go:31] will retry after 146.951232ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I1216 10:33:11.410618 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (5.037021705s)
I1216 10:33:11.412521 848599 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-109663 service yakd-dashboard -n yakd-dashboard
I1216 10:33:11.476073 848599 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I1216 10:33:11.476207 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:11.488433 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:11.488982 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:11.496693 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:11.558632 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I1216 10:33:11.682055 848599 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I1216 10:33:11.773243 848599 addons.go:234] Setting addon gcp-auth=true in "addons-109663"
I1216 10:33:11.773316 848599 host.go:66] Checking if "addons-109663" exists ...
I1216 10:33:11.773724 848599 cli_runner.go:164] Run: docker container inspect addons-109663 --format={{.State.Status}}
I1216 10:33:11.794622 848599 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I1216 10:33:11.794702 848599 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-109663
I1216 10:33:11.815157 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (3.527792528s)
I1216 10:33:11.815205 848599 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-109663"
I1216 10:33:11.816541 848599 out.go:177] * Verifying csi-hostpath-driver addon...
I1216 10:33:11.817906 848599 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33139 SSHKeyPath:/home/jenkins/minikube-integration/20107-840384/.minikube/machines/addons-109663/id_rsa Username:docker}
I1216 10:33:11.818272 848599 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I1216 10:33:11.876736 848599 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I1216 10:33:11.876761 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:11.988859 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:11.988992 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:12.321133 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:12.488213 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:12.488670 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:12.820552 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:12.988673 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:12.988814 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:13.183539 848599 node_ready.go:53] node "addons-109663" has status "Ready":"False"
I1216 10:33:13.321278 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:13.488254 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:13.488540 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:13.821374 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:13.988232 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:13.988468 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:14.319445 848599 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.760767877s)
I1216 10:33:14.319526 848599 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (2.524872285s)
I1216 10:33:14.321240 848599 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.4
I1216 10:33:14.321724 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:14.323613 848599 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.3
I1216 10:33:14.324764 848599 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I1216 10:33:14.324784 848599 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I1216 10:33:14.342010 848599 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I1216 10:33:14.342031 848599 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I1216 10:33:14.358225 848599 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I1216 10:33:14.358242 848599 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I1216 10:33:14.373588 848599 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.2/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I1216 10:33:14.489146 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:14.489195 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:14.688535 848599 addons.go:475] Verifying addon gcp-auth=true in "addons-109663"
I1216 10:33:14.689726 848599 out.go:177] * Verifying gcp-auth addon...
I1216 10:33:14.691774 848599 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I1216 10:33:14.693794 848599 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I1216 10:33:14.693814 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:14.821373 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:14.988073 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:14.988401 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:15.183828 848599 node_ready.go:53] node "addons-109663" has status "Ready":"False"
I1216 10:33:15.195110 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:15.321272 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:15.487994 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:15.488083 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:15.693975 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:15.821122 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:15.988405 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:15.989637 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:16.194046 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:16.321171 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:16.487823 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:16.488160 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:16.693727 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:16.821230 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:16.988099 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:16.988112 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:17.194466 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:17.320574 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:17.488270 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:17.488459 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:17.683559 848599 node_ready.go:53] node "addons-109663" has status "Ready":"False"
I1216 10:33:17.694310 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:17.821698 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:17.988850 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:17.989287 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:18.194610 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:18.320610 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:18.488098 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:18.488541 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:18.694691 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:18.820791 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:18.988425 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:18.988881 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:19.194979 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:19.321154 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:19.488238 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:19.488592 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:19.694150 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:19.821326 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:19.988036 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:19.988321 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:20.182762 848599 node_ready.go:53] node "addons-109663" has status "Ready":"False"
I1216 10:33:20.194584 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:20.322587 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:20.488030 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:20.488439 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:20.694318 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:20.821269 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:20.987972 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:20.988640 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:21.195178 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:21.321555 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:21.488262 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:21.488509 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:21.694792 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:21.821025 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:21.988166 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:21.988808 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:22.183545 848599 node_ready.go:49] node "addons-109663" has status "Ready":"True"
I1216 10:33:22.183575 848599 node_ready.go:38] duration metric: took 16.003041871s for node "addons-109663" to be "Ready" ...
I1216 10:33:22.183591 848599 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I1216 10:33:22.194312 848599 pod_ready.go:79] waiting up to 6m0s for pod "amd-gpu-device-plugin-nhj8x" in "kube-system" namespace to be "Ready" ...
I1216 10:33:22.197522 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:22.322250 848599 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I1216 10:33:22.322336 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:22.490226 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:22.490635 848599 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I1216 10:33:22.490660 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:22.696479 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:22.824283 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:22.991433 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:22.992361 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:23.195362 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:23.322483 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:23.489322 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:23.489683 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:23.695267 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:23.822532 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:23.989726 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:23.990631 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:24.195008 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:24.198627 848599 pod_ready.go:103] pod "amd-gpu-device-plugin-nhj8x" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:24.321616 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:24.489636 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:24.489791 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:24.695007 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:24.822843 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:24.988844 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:24.989122 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:25.195436 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:25.323096 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:25.488740 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:25.489022 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:25.694617 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:25.698720 848599 pod_ready.go:93] pod "amd-gpu-device-plugin-nhj8x" in "kube-system" namespace has status "Ready":"True"
I1216 10:33:25.698739 848599 pod_ready.go:82] duration metric: took 3.504402288s for pod "amd-gpu-device-plugin-nhj8x" in "kube-system" namespace to be "Ready" ...
I1216 10:33:25.698748 848599 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-ksv2k" in "kube-system" namespace to be "Ready" ...
I1216 10:33:25.702610 848599 pod_ready.go:93] pod "coredns-7c65d6cfc9-ksv2k" in "kube-system" namespace has status "Ready":"True"
I1216 10:33:25.702627 848599 pod_ready.go:82] duration metric: took 3.872629ms for pod "coredns-7c65d6cfc9-ksv2k" in "kube-system" namespace to be "Ready" ...
I1216 10:33:25.702644 848599 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-109663" in "kube-system" namespace to be "Ready" ...
I1216 10:33:25.706224 848599 pod_ready.go:93] pod "etcd-addons-109663" in "kube-system" namespace has status "Ready":"True"
I1216 10:33:25.706253 848599 pod_ready.go:82] duration metric: took 3.589378ms for pod "etcd-addons-109663" in "kube-system" namespace to be "Ready" ...
I1216 10:33:25.706269 848599 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-109663" in "kube-system" namespace to be "Ready" ...
I1216 10:33:25.709711 848599 pod_ready.go:93] pod "kube-apiserver-addons-109663" in "kube-system" namespace has status "Ready":"True"
I1216 10:33:25.709728 848599 pod_ready.go:82] duration metric: took 3.450709ms for pod "kube-apiserver-addons-109663" in "kube-system" namespace to be "Ready" ...
I1216 10:33:25.709736 848599 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-109663" in "kube-system" namespace to be "Ready" ...
I1216 10:33:25.713265 848599 pod_ready.go:93] pod "kube-controller-manager-addons-109663" in "kube-system" namespace has status "Ready":"True"
I1216 10:33:25.713281 848599 pod_ready.go:82] duration metric: took 3.538224ms for pod "kube-controller-manager-addons-109663" in "kube-system" namespace to be "Ready" ...
I1216 10:33:25.713292 848599 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-dw2js" in "kube-system" namespace to be "Ready" ...
I1216 10:33:25.822420 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:25.989042 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:25.989188 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:26.097285 848599 pod_ready.go:93] pod "kube-proxy-dw2js" in "kube-system" namespace has status "Ready":"True"
I1216 10:33:26.097307 848599 pod_ready.go:82] duration metric: took 384.009465ms for pod "kube-proxy-dw2js" in "kube-system" namespace to be "Ready" ...
I1216 10:33:26.097317 848599 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-109663" in "kube-system" namespace to be "Ready" ...
I1216 10:33:26.194937 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:26.322961 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:26.489581 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:26.489586 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:26.497405 848599 pod_ready.go:93] pod "kube-scheduler-addons-109663" in "kube-system" namespace has status "Ready":"True"
I1216 10:33:26.497429 848599 pod_ready.go:82] duration metric: took 400.104712ms for pod "kube-scheduler-addons-109663" in "kube-system" namespace to be "Ready" ...
I1216 10:33:26.497442 848599 pod_ready.go:79] waiting up to 6m0s for pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace to be "Ready" ...
I1216 10:33:26.696384 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:26.823165 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:26.989901 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:26.990164 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:27.195958 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:27.322795 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:27.489994 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:27.490526 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:27.695991 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:27.822606 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:27.989159 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:27.989525 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:28.195007 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:28.323133 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:28.489488 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:28.489846 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:28.502566 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:28.695362 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:28.823297 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:28.989059 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:28.989279 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:29.194687 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:29.375493 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:29.489995 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:29.493278 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:29.695153 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:29.823826 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:29.991148 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:29.991714 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:30.195680 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:30.322033 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:30.489421 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:30.489469 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:30.694988 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:30.876318 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:30.989582 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:30.990008 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:31.003003 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:31.195966 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:31.323096 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:31.489187 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:31.489815 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:31.695583 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:31.822994 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:31.988983 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:31.989237 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:32.195592 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:32.323277 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:32.488903 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:32.489392 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:32.696506 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:32.823243 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:32.988885 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:32.989148 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:33.196341 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:33.322674 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:33.488634 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:33.488667 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:33.502086 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:33.693815 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:33.821559 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:33.988429 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:33.988769 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:34.194152 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:34.325349 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:34.488396 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:34.488714 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:34.694574 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:34.876109 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:34.990189 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:34.990724 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:35.195194 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:35.375444 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:35.492191 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:35.493688 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:35.502341 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:35.695132 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:35.876051 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:35.990210 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:35.993798 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:36.195051 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:36.322207 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:36.489307 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:36.489410 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:36.696303 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:36.823455 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:36.989688 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:36.989711 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:37.195654 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:37.323150 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:37.489519 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:37.489577 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:37.502620 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:37.695819 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:37.823526 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:37.989517 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:37.989692 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:38.195490 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:38.323913 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:38.489512 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:38.489639 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:38.695009 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:38.823240 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:38.989637 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:38.989966 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:39.195863 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:39.322922 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:39.489532 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:39.489839 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:39.502745 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:39.695567 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:39.822774 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:39.989533 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:39.989845 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:40.195641 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:40.376146 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:40.490514 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:40.490606 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:40.696062 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:40.875447 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:40.989883 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:40.990067 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:41.196047 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:41.324554 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:41.489061 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:41.489773 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:41.502805 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:41.695780 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:41.823285 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:41.989357 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:41.989524 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:42.195514 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:42.323286 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:42.489536 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:42.489650 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:42.695923 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:42.823083 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:42.989403 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:42.989743 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:43.195428 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:43.374868 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:43.489367 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:43.489663 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:43.503985 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:43.696012 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:43.822523 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:43.989041 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:43.989507 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:44.196437 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:44.322827 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:44.489009 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:44.489985 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:44.695640 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:44.823416 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:44.989302 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I1216 10:33:44.989712 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:45.194931 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:45.322806 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:45.489197 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:45.489242 848599 kapi.go:107] duration metric: took 35.003353773s to wait for kubernetes.io/minikube-addons=registry ...
I1216 10:33:45.694246 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:45.822395 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:45.988886 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:46.003221 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:46.194491 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:46.322813 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:46.488662 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:46.694552 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:46.822549 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:46.989251 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:47.195148 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:47.322802 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:47.490015 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:47.694873 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:47.823284 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:47.989150 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:48.084009 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:48.195599 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:48.376370 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:48.489896 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:48.696319 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:48.876501 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:48.992663 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:49.195582 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:49.375953 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:49.490297 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:49.695610 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:49.823687 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:49.989632 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:50.195374 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:50.323555 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:50.489368 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:50.503787 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:50.695450 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:50.823129 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:50.988637 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:51.195270 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:51.323080 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:51.508782 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:51.695217 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:51.823494 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:51.989770 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:52.195617 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:52.321818 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:52.489915 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:52.696638 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:52.826492 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:52.988736 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:53.003486 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:53.195901 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:53.323388 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:53.490222 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:53.695771 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:53.824714 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:53.989114 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:54.195596 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:54.323421 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:54.488781 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:54.694998 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:54.822746 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:54.989975 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:55.195250 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:55.323393 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:55.489828 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:55.502636 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:55.695371 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:55.823548 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:55.988753 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:56.195311 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:56.322360 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:56.488475 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:56.695360 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:56.823160 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:56.988763 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:57.195557 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:57.323228 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:57.488760 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:57.502986 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:33:57.695786 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:57.822469 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:57.989745 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:58.194764 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:58.322146 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:58.489249 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:58.695078 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:58.822330 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:58.988720 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:59.195576 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:59.323259 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:33:59.490215 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:33:59.696072 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:33:59.823107 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:00.010677 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:00.012028 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:00.194618 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:34:00.322187 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:00.488296 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:00.695347 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:34:00.822633 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:00.989168 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:01.194868 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:34:01.322119 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:01.489010 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:01.695366 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:34:01.823165 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:01.988799 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:02.194750 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I1216 10:34:02.322202 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:02.488979 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:02.502416 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:02.695830 848599 kapi.go:107] duration metric: took 48.004052123s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I1216 10:34:02.697415 848599 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-109663 cluster.
I1216 10:34:02.698542 848599 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I1216 10:34:02.699693 848599 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I1216 10:34:02.874592 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:02.989784 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:03.322719 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:03.489551 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:03.823450 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:03.989440 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:04.324087 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:04.489445 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:04.502644 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:04.822353 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:04.989069 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:05.323880 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:05.512296 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:05.875523 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:05.990166 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:06.397131 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:06.489855 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:06.574453 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:06.877071 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:06.989900 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:07.376293 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:07.494377 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:07.878721 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:07.988468 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:08.323271 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:08.489196 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:08.823385 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:08.988831 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:09.003062 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:09.323192 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:09.489553 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:09.822684 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:09.989937 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:10.323395 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:10.489825 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:10.823045 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:10.988842 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:11.003275 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:11.322851 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:11.489655 848599 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I1216 10:34:11.823870 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:11.989881 848599 kapi.go:107] duration metric: took 1m1.505001565s to wait for app.kubernetes.io/name=ingress-nginx ...
I1216 10:34:12.322605 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:12.876290 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:13.003959 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:13.324089 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:13.823333 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:14.322846 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:14.822531 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:15.323702 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:15.503382 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:15.822069 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:16.322918 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:16.822182 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:17.322594 848599 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I1216 10:34:17.823222 848599 kapi.go:107] duration metric: took 1m6.004947421s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I1216 10:34:17.824726 848599 out.go:177] * Enabled addons: storage-provisioner, amd-gpu-device-plugin, cloud-spanner, inspektor-gadget, ingress-dns, nvidia-device-plugin, metrics-server, default-storageclass, yakd, volumesnapshots, registry, gcp-auth, ingress, csi-hostpath-driver
I1216 10:34:17.825819 848599 addons.go:510] duration metric: took 1m13.736912479s for enable addons: enabled=[storage-provisioner amd-gpu-device-plugin cloud-spanner inspektor-gadget ingress-dns nvidia-device-plugin metrics-server default-storageclass yakd volumesnapshots registry gcp-auth ingress csi-hostpath-driver]
I1216 10:34:18.003373 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:20.502137 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:22.502669 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:24.503206 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:26.503344 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:28.620772 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:31.003959 848599 pod_ready.go:103] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"False"
I1216 10:34:33.002781 848599 pod_ready.go:93] pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace has status "Ready":"True"
I1216 10:34:33.002804 848599 pod_ready.go:82] duration metric: took 1m6.505353818s for pod "metrics-server-84c5f94fbc-z8rzz" in "kube-system" namespace to be "Ready" ...
I1216 10:34:33.002816 848599 pod_ready.go:79] waiting up to 6m0s for pod "nvidia-device-plugin-daemonset-k4znm" in "kube-system" namespace to be "Ready" ...
I1216 10:34:33.007179 848599 pod_ready.go:93] pod "nvidia-device-plugin-daemonset-k4znm" in "kube-system" namespace has status "Ready":"True"
I1216 10:34:33.007200 848599 pod_ready.go:82] duration metric: took 4.376449ms for pod "nvidia-device-plugin-daemonset-k4znm" in "kube-system" namespace to be "Ready" ...
I1216 10:34:33.007222 848599 pod_ready.go:39] duration metric: took 1m10.823613152s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I1216 10:34:33.007246 848599 api_server.go:52] waiting for apiserver process to appear ...
I1216 10:34:33.007317 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-apiserver Namespaces:[]}
I1216 10:34:33.007419 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-apiserver
I1216 10:34:33.041826 848599 cri.go:89] found id: "c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804"
I1216 10:34:33.041843 848599 cri.go:89] found id: ""
I1216 10:34:33.041852 848599 logs.go:282] 1 containers: [c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804]
I1216 10:34:33.041893 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:33.045200 848599 cri.go:54] listing CRI containers in root : {State:all Name:etcd Namespaces:[]}
I1216 10:34:33.045244 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=etcd
I1216 10:34:33.078373 848599 cri.go:89] found id: "93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6"
I1216 10:34:33.078390 848599 cri.go:89] found id: ""
I1216 10:34:33.078398 848599 logs.go:282] 1 containers: [93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6]
I1216 10:34:33.078432 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:33.081776 848599 cri.go:54] listing CRI containers in root : {State:all Name:coredns Namespaces:[]}
I1216 10:34:33.081822 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=coredns
I1216 10:34:33.113665 848599 cri.go:89] found id: "d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190"
I1216 10:34:33.113684 848599 cri.go:89] found id: ""
I1216 10:34:33.113692 848599 logs.go:282] 1 containers: [d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190]
I1216 10:34:33.113726 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:33.116711 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-scheduler Namespaces:[]}
I1216 10:34:33.116773 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-scheduler
I1216 10:34:33.149097 848599 cri.go:89] found id: "c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53"
I1216 10:34:33.149116 848599 cri.go:89] found id: ""
I1216 10:34:33.149129 848599 logs.go:282] 1 containers: [c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53]
I1216 10:34:33.149163 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:33.152109 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-proxy Namespaces:[]}
I1216 10:34:33.152155 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-proxy
I1216 10:34:33.182865 848599 cri.go:89] found id: "c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc"
I1216 10:34:33.182884 848599 cri.go:89] found id: ""
I1216 10:34:33.182894 848599 logs.go:282] 1 containers: [c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc]
I1216 10:34:33.182927 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:33.185812 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-controller-manager Namespaces:[]}
I1216 10:34:33.185877 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-controller-manager
I1216 10:34:33.217210 848599 cri.go:89] found id: "bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50"
I1216 10:34:33.217232 848599 cri.go:89] found id: ""
I1216 10:34:33.217244 848599 logs.go:282] 1 containers: [bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50]
I1216 10:34:33.217278 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:33.220246 848599 cri.go:54] listing CRI containers in root : {State:all Name:kindnet Namespaces:[]}
I1216 10:34:33.220314 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kindnet
I1216 10:34:33.252320 848599 cri.go:89] found id: "9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9"
I1216 10:34:33.252355 848599 cri.go:89] found id: ""
I1216 10:34:33.252367 848599 logs.go:282] 1 containers: [9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9]
I1216 10:34:33.252411 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:33.255333 848599 logs.go:123] Gathering logs for kubelet ...
I1216 10:34:33.255361 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo journalctl -u kubelet -n 400"
I1216 10:34:33.334752 848599 logs.go:123] Gathering logs for kube-apiserver [c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804] ...
I1216 10:34:33.334782 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804"
I1216 10:34:33.377060 848599 logs.go:123] Gathering logs for kube-proxy [c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc] ...
I1216 10:34:33.377081 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc"
I1216 10:34:33.409466 848599 logs.go:123] Gathering logs for kube-controller-manager [bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50] ...
I1216 10:34:33.409490 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50"
I1216 10:34:33.463839 848599 logs.go:123] Gathering logs for CRI-O ...
I1216 10:34:33.463865 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo journalctl -u crio -n 400"
I1216 10:34:33.536966 848599 logs.go:123] Gathering logs for container status ...
I1216 10:34:33.536995 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo `which crictl || echo crictl` ps -a || sudo docker ps -a"
I1216 10:34:33.576693 848599 logs.go:123] Gathering logs for dmesg ...
I1216 10:34:33.576718 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo dmesg -PH -L=never --level warn,err,crit,alert,emerg | tail -n 400"
I1216 10:34:33.602210 848599 logs.go:123] Gathering logs for describe nodes ...
I1216 10:34:33.602235 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.2/kubectl describe nodes --kubeconfig=/var/lib/minikube/kubeconfig"
I1216 10:34:33.698067 848599 logs.go:123] Gathering logs for etcd [93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6] ...
I1216 10:34:33.698107 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6"
I1216 10:34:33.748695 848599 logs.go:123] Gathering logs for coredns [d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190] ...
I1216 10:34:33.748723 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190"
I1216 10:34:33.802810 848599 logs.go:123] Gathering logs for kube-scheduler [c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53] ...
I1216 10:34:33.802846 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53"
I1216 10:34:33.841795 848599 logs.go:123] Gathering logs for kindnet [9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9] ...
I1216 10:34:33.841823 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9"
I1216 10:34:36.373886 848599 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I1216 10:34:36.387690 848599 api_server.go:72] duration metric: took 1m32.298817651s to wait for apiserver process to appear ...
I1216 10:34:36.387721 848599 api_server.go:88] waiting for apiserver healthz status ...
I1216 10:34:36.387772 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-apiserver Namespaces:[]}
I1216 10:34:36.387841 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-apiserver
I1216 10:34:36.421031 848599 cri.go:89] found id: "c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804"
I1216 10:34:36.421062 848599 cri.go:89] found id: ""
I1216 10:34:36.421077 848599 logs.go:282] 1 containers: [c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804]
I1216 10:34:36.421138 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:36.424373 848599 cri.go:54] listing CRI containers in root : {State:all Name:etcd Namespaces:[]}
I1216 10:34:36.424428 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=etcd
I1216 10:34:36.456413 848599 cri.go:89] found id: "93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6"
I1216 10:34:36.456434 848599 cri.go:89] found id: ""
I1216 10:34:36.456445 848599 logs.go:282] 1 containers: [93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6]
I1216 10:34:36.456495 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:36.459492 848599 cri.go:54] listing CRI containers in root : {State:all Name:coredns Namespaces:[]}
I1216 10:34:36.459554 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=coredns
I1216 10:34:36.491350 848599 cri.go:89] found id: "d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190"
I1216 10:34:36.491370 848599 cri.go:89] found id: ""
I1216 10:34:36.491379 848599 logs.go:282] 1 containers: [d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190]
I1216 10:34:36.491420 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:36.494403 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-scheduler Namespaces:[]}
I1216 10:34:36.494454 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-scheduler
I1216 10:34:36.526671 848599 cri.go:89] found id: "c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53"
I1216 10:34:36.526688 848599 cri.go:89] found id: ""
I1216 10:34:36.526695 848599 logs.go:282] 1 containers: [c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53]
I1216 10:34:36.526735 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:36.529636 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-proxy Namespaces:[]}
I1216 10:34:36.529688 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-proxy
I1216 10:34:36.563198 848599 cri.go:89] found id: "c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc"
I1216 10:34:36.563217 848599 cri.go:89] found id: ""
I1216 10:34:36.563227 848599 logs.go:282] 1 containers: [c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc]
I1216 10:34:36.563283 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:36.566202 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-controller-manager Namespaces:[]}
I1216 10:34:36.566256 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-controller-manager
I1216 10:34:36.598334 848599 cri.go:89] found id: "bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50"
I1216 10:34:36.598353 848599 cri.go:89] found id: ""
I1216 10:34:36.598361 848599 logs.go:282] 1 containers: [bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50]
I1216 10:34:36.598413 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:36.601335 848599 cri.go:54] listing CRI containers in root : {State:all Name:kindnet Namespaces:[]}
I1216 10:34:36.601404 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kindnet
I1216 10:34:36.634180 848599 cri.go:89] found id: "9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9"
I1216 10:34:36.634195 848599 cri.go:89] found id: ""
I1216 10:34:36.634203 848599 logs.go:282] 1 containers: [9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9]
I1216 10:34:36.634250 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:36.637167 848599 logs.go:123] Gathering logs for kube-apiserver [c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804] ...
I1216 10:34:36.637191 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804"
I1216 10:34:36.680397 848599 logs.go:123] Gathering logs for kube-scheduler [c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53] ...
I1216 10:34:36.680421 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53"
I1216 10:34:36.717036 848599 logs.go:123] Gathering logs for kube-controller-manager [bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50] ...
I1216 10:34:36.717062 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50"
I1216 10:34:36.771623 848599 logs.go:123] Gathering logs for CRI-O ...
I1216 10:34:36.771648 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo journalctl -u crio -n 400"
I1216 10:34:36.848400 848599 logs.go:123] Gathering logs for container status ...
I1216 10:34:36.848426 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo `which crictl || echo crictl` ps -a || sudo docker ps -a"
I1216 10:34:36.890499 848599 logs.go:123] Gathering logs for describe nodes ...
I1216 10:34:36.890524 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.2/kubectl describe nodes --kubeconfig=/var/lib/minikube/kubeconfig"
I1216 10:34:36.984658 848599 logs.go:123] Gathering logs for dmesg ...
I1216 10:34:36.984683 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo dmesg -PH -L=never --level warn,err,crit,alert,emerg | tail -n 400"
I1216 10:34:37.010767 848599 logs.go:123] Gathering logs for etcd [93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6] ...
I1216 10:34:37.010795 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6"
I1216 10:34:37.058997 848599 logs.go:123] Gathering logs for coredns [d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190] ...
I1216 10:34:37.059021 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190"
I1216 10:34:37.109478 848599 logs.go:123] Gathering logs for kube-proxy [c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc] ...
I1216 10:34:37.109513 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc"
I1216 10:34:37.141261 848599 logs.go:123] Gathering logs for kindnet [9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9] ...
I1216 10:34:37.141284 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9"
I1216 10:34:37.173779 848599 logs.go:123] Gathering logs for kubelet ...
I1216 10:34:37.173857 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo journalctl -u kubelet -n 400"
I1216 10:34:39.755150 848599 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I1216 10:34:39.758789 848599 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I1216 10:34:39.759732 848599 api_server.go:141] control plane version: v1.31.2
I1216 10:34:39.759759 848599 api_server.go:131] duration metric: took 3.372030509s to wait for apiserver health ...
I1216 10:34:39.759767 848599 system_pods.go:43] waiting for kube-system pods to appear ...
I1216 10:34:39.759796 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-apiserver Namespaces:[]}
I1216 10:34:39.759850 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-apiserver
I1216 10:34:39.795031 848599 cri.go:89] found id: "c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804"
I1216 10:34:39.795051 848599 cri.go:89] found id: ""
I1216 10:34:39.795060 848599 logs.go:282] 1 containers: [c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804]
I1216 10:34:39.795104 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:39.798358 848599 cri.go:54] listing CRI containers in root : {State:all Name:etcd Namespaces:[]}
I1216 10:34:39.798435 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=etcd
I1216 10:34:39.831890 848599 cri.go:89] found id: "93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6"
I1216 10:34:39.831906 848599 cri.go:89] found id: ""
I1216 10:34:39.831913 848599 logs.go:282] 1 containers: [93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6]
I1216 10:34:39.831951 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:39.834968 848599 cri.go:54] listing CRI containers in root : {State:all Name:coredns Namespaces:[]}
I1216 10:34:39.835037 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=coredns
I1216 10:34:39.866579 848599 cri.go:89] found id: "d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190"
I1216 10:34:39.866602 848599 cri.go:89] found id: ""
I1216 10:34:39.866613 848599 logs.go:282] 1 containers: [d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190]
I1216 10:34:39.866647 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:39.869695 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-scheduler Namespaces:[]}
I1216 10:34:39.869763 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-scheduler
I1216 10:34:39.901933 848599 cri.go:89] found id: "c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53"
I1216 10:34:39.901954 848599 cri.go:89] found id: ""
I1216 10:34:39.901966 848599 logs.go:282] 1 containers: [c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53]
I1216 10:34:39.902014 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:39.905112 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-proxy Namespaces:[]}
I1216 10:34:39.905174 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-proxy
I1216 10:34:39.938572 848599 cri.go:89] found id: "c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc"
I1216 10:34:39.938590 848599 cri.go:89] found id: ""
I1216 10:34:39.938598 848599 logs.go:282] 1 containers: [c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc]
I1216 10:34:39.938648 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:39.941675 848599 cri.go:54] listing CRI containers in root : {State:all Name:kube-controller-manager Namespaces:[]}
I1216 10:34:39.941738 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kube-controller-manager
I1216 10:34:39.974011 848599 cri.go:89] found id: "bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50"
I1216 10:34:39.974033 848599 cri.go:89] found id: ""
I1216 10:34:39.974043 848599 logs.go:282] 1 containers: [bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50]
I1216 10:34:39.974092 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:39.977679 848599 cri.go:54] listing CRI containers in root : {State:all Name:kindnet Namespaces:[]}
I1216 10:34:39.977725 848599 ssh_runner.go:195] Run: sudo crictl ps -a --quiet --name=kindnet
I1216 10:34:40.011518 848599 cri.go:89] found id: "9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9"
I1216 10:34:40.011539 848599 cri.go:89] found id: ""
I1216 10:34:40.011547 848599 logs.go:282] 1 containers: [9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9]
I1216 10:34:40.011598 848599 ssh_runner.go:195] Run: which crictl
I1216 10:34:40.014781 848599 logs.go:123] Gathering logs for kubelet ...
I1216 10:34:40.014805 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo journalctl -u kubelet -n 400"
I1216 10:34:40.093024 848599 logs.go:123] Gathering logs for kube-proxy [c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc] ...
I1216 10:34:40.093048 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc"
I1216 10:34:40.125022 848599 logs.go:123] Gathering logs for container status ...
I1216 10:34:40.125045 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo `which crictl || echo crictl` ps -a || sudo docker ps -a"
I1216 10:34:40.167202 848599 logs.go:123] Gathering logs for etcd [93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6] ...
I1216 10:34:40.167230 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6"
I1216 10:34:40.217550 848599 logs.go:123] Gathering logs for coredns [d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190] ...
I1216 10:34:40.217579 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190"
I1216 10:34:40.271787 848599 logs.go:123] Gathering logs for kube-scheduler [c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53] ...
I1216 10:34:40.271829 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53"
I1216 10:34:40.308809 848599 logs.go:123] Gathering logs for kube-controller-manager [bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50] ...
I1216 10:34:40.308835 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50"
I1216 10:34:40.363908 848599 logs.go:123] Gathering logs for kindnet [9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9] ...
I1216 10:34:40.363934 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9"
I1216 10:34:40.396438 848599 logs.go:123] Gathering logs for dmesg ...
I1216 10:34:40.396463 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo dmesg -PH -L=never --level warn,err,crit,alert,emerg | tail -n 400"
I1216 10:34:40.424000 848599 logs.go:123] Gathering logs for describe nodes ...
I1216 10:34:40.424023 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.2/kubectl describe nodes --kubeconfig=/var/lib/minikube/kubeconfig"
I1216 10:34:40.521844 848599 logs.go:123] Gathering logs for kube-apiserver [c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804] ...
I1216 10:34:40.521880 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo /usr/bin/crictl logs --tail 400 c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804"
I1216 10:34:40.565349 848599 logs.go:123] Gathering logs for CRI-O ...
I1216 10:34:40.565379 848599 ssh_runner.go:195] Run: /bin/bash -c "sudo journalctl -u crio -n 400"
I1216 10:34:43.152582 848599 system_pods.go:59] 19 kube-system pods found
I1216 10:34:43.152628 848599 system_pods.go:61] "amd-gpu-device-plugin-nhj8x" [483a0808-3e15-4de2-b48a-ecfa43394c55] Running
I1216 10:34:43.152639 848599 system_pods.go:61] "coredns-7c65d6cfc9-ksv2k" [b31289fc-3ff8-4af0-a5d2-a88dace5589c] Running
I1216 10:34:43.152645 848599 system_pods.go:61] "csi-hostpath-attacher-0" [9089b466-c717-4755-bf51-2740aecfaeb6] Running
I1216 10:34:43.152650 848599 system_pods.go:61] "csi-hostpath-resizer-0" [963124d9-8e43-4fb9-a011-05c542d2fb50] Running
I1216 10:34:43.152655 848599 system_pods.go:61] "csi-hostpathplugin-7826x" [856ef16b-5b68-404c-8df4-558dc73fe76b] Running
I1216 10:34:43.152660 848599 system_pods.go:61] "etcd-addons-109663" [9789d971-2bea-46bf-872e-e096afce5cb0] Running
I1216 10:34:43.152666 848599 system_pods.go:61] "kindnet-sn2ww" [1c8f1cfd-5f82-439c-b6f7-b654f855b517] Running
I1216 10:34:43.152672 848599 system_pods.go:61] "kube-apiserver-addons-109663" [4e04829b-d42e-4de8-be6a-0ec8196b7c28] Running
I1216 10:34:43.152678 848599 system_pods.go:61] "kube-controller-manager-addons-109663" [c5a39a90-0604-42e4-bdc4-d4b9ab6f6df5] Running
I1216 10:34:43.152687 848599 system_pods.go:61] "kube-ingress-dns-minikube" [a0ba89f2-e8b1-498e-ab03-dd8a5e50c176] Running
I1216 10:34:43.152694 848599 system_pods.go:61] "kube-proxy-dw2js" [82afbc0e-6ed6-4a7a-8721-d77176570525] Running
I1216 10:34:43.152703 848599 system_pods.go:61] "kube-scheduler-addons-109663" [018079f5-5c1a-4a2c-8845-8adfc665ce77] Running
I1216 10:34:43.152709 848599 system_pods.go:61] "metrics-server-84c5f94fbc-z8rzz" [0c4013ee-0e9e-4bf6-aff8-752bb76b1c0c] Running
I1216 10:34:43.152719 848599 system_pods.go:61] "nvidia-device-plugin-daemonset-k4znm" [94be2280-9ef7-49a1-aed5-ae48c7b50056] Running
I1216 10:34:43.152725 848599 system_pods.go:61] "registry-5cc95cd69-rkb22" [9148bfd2-bdfd-42f6-9b6e-f2cb29de4e1e] Running
I1216 10:34:43.152731 848599 system_pods.go:61] "registry-proxy-w5gg9" [5d79e061-c009-4296-adaf-94ec1a94ed36] Running
I1216 10:34:43.152737 848599 system_pods.go:61] "snapshot-controller-56fcc65765-8skj8" [29ea6b74-8543-4d6d-a9f0-8476aaef7f19] Running
I1216 10:34:43.152744 848599 system_pods.go:61] "snapshot-controller-56fcc65765-rb9fx" [62bd9cad-e4a7-474c-9ce0-bb38412ded35] Running
I1216 10:34:43.152752 848599 system_pods.go:61] "storage-provisioner" [f6eecac1-47ca-4d5e-8014-bbb9f35f7213] Running
I1216 10:34:43.152764 848599 system_pods.go:74] duration metric: took 3.392988839s to wait for pod list to return data ...
I1216 10:34:43.152779 848599 default_sa.go:34] waiting for default service account to be created ...
I1216 10:34:43.154908 848599 default_sa.go:45] found service account: "default"
I1216 10:34:43.154931 848599 default_sa.go:55] duration metric: took 2.143478ms for default service account to be created ...
I1216 10:34:43.154942 848599 system_pods.go:116] waiting for k8s-apps to be running ...
I1216 10:34:43.164127 848599 system_pods.go:86] 19 kube-system pods found
I1216 10:34:43.164152 848599 system_pods.go:89] "amd-gpu-device-plugin-nhj8x" [483a0808-3e15-4de2-b48a-ecfa43394c55] Running
I1216 10:34:43.164158 848599 system_pods.go:89] "coredns-7c65d6cfc9-ksv2k" [b31289fc-3ff8-4af0-a5d2-a88dace5589c] Running
I1216 10:34:43.164162 848599 system_pods.go:89] "csi-hostpath-attacher-0" [9089b466-c717-4755-bf51-2740aecfaeb6] Running
I1216 10:34:43.164166 848599 system_pods.go:89] "csi-hostpath-resizer-0" [963124d9-8e43-4fb9-a011-05c542d2fb50] Running
I1216 10:34:43.164170 848599 system_pods.go:89] "csi-hostpathplugin-7826x" [856ef16b-5b68-404c-8df4-558dc73fe76b] Running
I1216 10:34:43.164173 848599 system_pods.go:89] "etcd-addons-109663" [9789d971-2bea-46bf-872e-e096afce5cb0] Running
I1216 10:34:43.164176 848599 system_pods.go:89] "kindnet-sn2ww" [1c8f1cfd-5f82-439c-b6f7-b654f855b517] Running
I1216 10:34:43.164180 848599 system_pods.go:89] "kube-apiserver-addons-109663" [4e04829b-d42e-4de8-be6a-0ec8196b7c28] Running
I1216 10:34:43.164184 848599 system_pods.go:89] "kube-controller-manager-addons-109663" [c5a39a90-0604-42e4-bdc4-d4b9ab6f6df5] Running
I1216 10:34:43.164189 848599 system_pods.go:89] "kube-ingress-dns-minikube" [a0ba89f2-e8b1-498e-ab03-dd8a5e50c176] Running
I1216 10:34:43.164195 848599 system_pods.go:89] "kube-proxy-dw2js" [82afbc0e-6ed6-4a7a-8721-d77176570525] Running
I1216 10:34:43.164199 848599 system_pods.go:89] "kube-scheduler-addons-109663" [018079f5-5c1a-4a2c-8845-8adfc665ce77] Running
I1216 10:34:43.164203 848599 system_pods.go:89] "metrics-server-84c5f94fbc-z8rzz" [0c4013ee-0e9e-4bf6-aff8-752bb76b1c0c] Running
I1216 10:34:43.164208 848599 system_pods.go:89] "nvidia-device-plugin-daemonset-k4znm" [94be2280-9ef7-49a1-aed5-ae48c7b50056] Running
I1216 10:34:43.164220 848599 system_pods.go:89] "registry-5cc95cd69-rkb22" [9148bfd2-bdfd-42f6-9b6e-f2cb29de4e1e] Running
I1216 10:34:43.164223 848599 system_pods.go:89] "registry-proxy-w5gg9" [5d79e061-c009-4296-adaf-94ec1a94ed36] Running
I1216 10:34:43.164228 848599 system_pods.go:89] "snapshot-controller-56fcc65765-8skj8" [29ea6b74-8543-4d6d-a9f0-8476aaef7f19] Running
I1216 10:34:43.164234 848599 system_pods.go:89] "snapshot-controller-56fcc65765-rb9fx" [62bd9cad-e4a7-474c-9ce0-bb38412ded35] Running
I1216 10:34:43.164237 848599 system_pods.go:89] "storage-provisioner" [f6eecac1-47ca-4d5e-8014-bbb9f35f7213] Running
I1216 10:34:43.164244 848599 system_pods.go:126] duration metric: took 9.295549ms to wait for k8s-apps to be running ...
I1216 10:34:43.164253 848599 system_svc.go:44] waiting for kubelet service to be running ....
I1216 10:34:43.164295 848599 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I1216 10:34:43.175918 848599 system_svc.go:56] duration metric: took 11.65853ms WaitForService to wait for kubelet
I1216 10:34:43.175940 848599 kubeadm.go:582] duration metric: took 1m39.087076667s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I1216 10:34:43.175962 848599 node_conditions.go:102] verifying NodePressure condition ...
I1216 10:34:43.178532 848599 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I1216 10:34:43.178559 848599 node_conditions.go:123] node cpu capacity is 8
I1216 10:34:43.178575 848599 node_conditions.go:105] duration metric: took 2.605732ms to run NodePressure ...
I1216 10:34:43.178594 848599 start.go:241] waiting for startup goroutines ...
I1216 10:34:43.178609 848599 start.go:246] waiting for cluster config update ...
I1216 10:34:43.178631 848599 start.go:255] writing updated cluster config ...
I1216 10:34:43.178953 848599 ssh_runner.go:195] Run: rm -f paused
I1216 10:34:43.230691 848599 start.go:600] kubectl: 1.32.0, cluster: 1.31.2 (minor skew: 1)
I1216 10:34:43.232683 848599 out.go:177] * Done! kubectl is now configured to use "addons-109663" cluster and "default" namespace by default
==> CRI-O <==
Dec 16 10:36:58 addons-109663 crio[1041]: time="2024-12-16 10:36:58.273779077Z" level=info msg="Removed pod sandbox: 8bc46f18bedb31da572fc9f1c74f40e93ff832db94b07b6e643b6bacf285ad88" id=7ae85b51-0525-4420-a7e3-a4418fc23787 name=/runtime.v1.RuntimeService/RemovePodSandbox
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.696379738Z" level=info msg="Running pod sandbox: default/hello-world-app-55bf9c44b4-br7qj/POD" id=03f4d38f-128f-4049-8223-559b014b676e name=/runtime.v1.RuntimeService/RunPodSandbox
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.696446086Z" level=warning msg="Allowed annotations are specified for workload []"
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.714124632Z" level=info msg="Got pod network &{Name:hello-world-app-55bf9c44b4-br7qj Namespace:default ID:fa2caf81dd3dd6b79e8274a68d3ea74e6c9e0dacfe5922ddb2764dc2c0eb52b7 UID:6046c7ab-0532-4ad2-907c-cbe45f15d836 NetNS:/var/run/netns/0429debe-237e-4f9e-8adb-2edda12918e1 Networks:[] RuntimeConfig:map[kindnet:{IP: MAC: PortMappings:[] Bandwidth:<nil> IpRanges:[]}] Aliases:map[]}"
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.714158222Z" level=info msg="Adding pod default_hello-world-app-55bf9c44b4-br7qj to CNI network \"kindnet\" (type=ptp)"
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.725472994Z" level=info msg="Got pod network &{Name:hello-world-app-55bf9c44b4-br7qj Namespace:default ID:fa2caf81dd3dd6b79e8274a68d3ea74e6c9e0dacfe5922ddb2764dc2c0eb52b7 UID:6046c7ab-0532-4ad2-907c-cbe45f15d836 NetNS:/var/run/netns/0429debe-237e-4f9e-8adb-2edda12918e1 Networks:[] RuntimeConfig:map[kindnet:{IP: MAC: PortMappings:[] Bandwidth:<nil> IpRanges:[]}] Aliases:map[]}"
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.725657315Z" level=info msg="Checking pod default_hello-world-app-55bf9c44b4-br7qj for CNI network kindnet (type=ptp)"
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.729386244Z" level=info msg="Ran pod sandbox fa2caf81dd3dd6b79e8274a68d3ea74e6c9e0dacfe5922ddb2764dc2c0eb52b7 with infra container: default/hello-world-app-55bf9c44b4-br7qj/POD" id=03f4d38f-128f-4049-8223-559b014b676e name=/runtime.v1.RuntimeService/RunPodSandbox
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.730926434Z" level=info msg="Checking image status: docker.io/kicbase/echo-server:1.0" id=9aa4f830-863e-48de-975f-8fe81a3b75af name=/runtime.v1.ImageService/ImageStatus
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.731209976Z" level=info msg="Image docker.io/kicbase/echo-server:1.0 not found" id=9aa4f830-863e-48de-975f-8fe81a3b75af name=/runtime.v1.ImageService/ImageStatus
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.772415488Z" level=info msg="Pulling image: docker.io/kicbase/echo-server:1.0" id=71a5bd68-1af8-40f8-a060-2d05e2acdaaa name=/runtime.v1.ImageService/PullImage
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.776194669Z" level=info msg="Trying to access \"docker.io/kicbase/echo-server:1.0\""
Dec 16 10:37:42 addons-109663 crio[1041]: time="2024-12-16 10:37:42.921108659Z" level=info msg="Trying to access \"docker.io/kicbase/echo-server:1.0\""
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.322622647Z" level=info msg="Pulled image: docker.io/kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6" id=71a5bd68-1af8-40f8-a060-2d05e2acdaaa name=/runtime.v1.ImageService/PullImage
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.323238994Z" level=info msg="Checking image status: docker.io/kicbase/echo-server:1.0" id=9e5124c2-ea9a-4f5f-9959-5c7614c82eb9 name=/runtime.v1.ImageService/ImageStatus
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.324452595Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9056ab77afb8e18e04303f11000a9d31b3f16b74c59475b899ae1b342d328d30,RepoTags:[docker.io/kicbase/echo-server:1.0],RepoDigests:[docker.io/kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 docker.io/kicbase/echo-server@sha256:a82eba7887a40ecae558433f34225b2611dc77f982ce05b1ddb9b282b780fc86],Size_:4944818,Uid:nil,Username:,Spec:nil,},Info:map[string]string{},}" id=9e5124c2-ea9a-4f5f-9959-5c7614c82eb9 name=/runtime.v1.ImageService/ImageStatus
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.326193580Z" level=info msg="Checking image status: docker.io/kicbase/echo-server:1.0" id=ad445986-08aa-4adf-ac4c-da8e4ded7b21 name=/runtime.v1.ImageService/ImageStatus
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.327176561Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9056ab77afb8e18e04303f11000a9d31b3f16b74c59475b899ae1b342d328d30,RepoTags:[docker.io/kicbase/echo-server:1.0],RepoDigests:[docker.io/kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 docker.io/kicbase/echo-server@sha256:a82eba7887a40ecae558433f34225b2611dc77f982ce05b1ddb9b282b780fc86],Size_:4944818,Uid:nil,Username:,Spec:nil,},Info:map[string]string{},}" id=ad445986-08aa-4adf-ac4c-da8e4ded7b21 name=/runtime.v1.ImageService/ImageStatus
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.328052355Z" level=info msg="Creating container: default/hello-world-app-55bf9c44b4-br7qj/hello-world-app" id=7733217e-2acf-4800-bf1f-fedba486f42c name=/runtime.v1.RuntimeService/CreateContainer
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.328158984Z" level=warning msg="Allowed annotations are specified for workload []"
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.341364639Z" level=warning msg="Failed to open /etc/passwd: open /var/lib/containers/storage/overlay/03ae318c7ff1cec3065d7c2fd68dad8cd3ebec157e59aa878c20ac29e21fcf22/merged/etc/passwd: no such file or directory"
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.341394337Z" level=warning msg="Failed to open /etc/group: open /var/lib/containers/storage/overlay/03ae318c7ff1cec3065d7c2fd68dad8cd3ebec157e59aa878c20ac29e21fcf22/merged/etc/group: no such file or directory"
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.378072255Z" level=info msg="Created container b6ac96fbd2ff784e0fe00a1f772ffebf5fb4eaac1854f9781ffc6384fb2d4a71: default/hello-world-app-55bf9c44b4-br7qj/hello-world-app" id=7733217e-2acf-4800-bf1f-fedba486f42c name=/runtime.v1.RuntimeService/CreateContainer
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.378607045Z" level=info msg="Starting container: b6ac96fbd2ff784e0fe00a1f772ffebf5fb4eaac1854f9781ffc6384fb2d4a71" id=8bcdaa37-b955-4f43-852b-712c73fc58a7 name=/runtime.v1.RuntimeService/StartContainer
Dec 16 10:37:43 addons-109663 crio[1041]: time="2024-12-16 10:37:43.384254350Z" level=info msg="Started container" PID=11076 containerID=b6ac96fbd2ff784e0fe00a1f772ffebf5fb4eaac1854f9781ffc6384fb2d4a71 description=default/hello-world-app-55bf9c44b4-br7qj/hello-world-app id=8bcdaa37-b955-4f43-852b-712c73fc58a7 name=/runtime.v1.RuntimeService/StartContainer sandboxID=fa2caf81dd3dd6b79e8274a68d3ea74e6c9e0dacfe5922ddb2764dc2c0eb52b7
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
b6ac96fbd2ff7 docker.io/kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 Less than a second ago Running hello-world-app 0 fa2caf81dd3dd hello-world-app-55bf9c44b4-br7qj
d384a65188bd6 docker.io/library/nginx@sha256:41523187cf7d7a2f2677a80609d9caa14388bf5c1fbca9c410ba3de602aaaab4 2 minutes ago Running nginx 0 85a1116f8171d nginx
e0df8328f54c0 gcr.io/k8s-minikube/busybox@sha256:2d03e6ceeb99250061dd110530b0ece7998cd84121f952adef120ea7c5a6f00e 2 minutes ago Running busybox 0 cab241fcb05db busybox
14ccae418ccad registry.k8s.io/ingress-nginx/controller@sha256:62b61c42ec8dd877b85c0aa24c4744ce44d274bc16cc5d2364edfe67964ba55b 3 minutes ago Running controller 0 96c18162d52c8 ingress-nginx-controller-5f85ff4588-5q5qg
58e1003acbde1 gcr.io/k8s-minikube/minikube-ingress-dns@sha256:07c8f5b205a3f8971bfc6d460978ae00de35f17e5d5392b1de8de02356f85dab 4 minutes ago Running minikube-ingress-dns 0 0b71d4cf0076c kube-ingress-dns-minikube
35888134a837c registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a9f03b34a3cbfbb26d103a14046ab2c5130a80c3d69d526ff8063d2b37b9fd3f 4 minutes ago Exited patch 0 282892f631a51 ingress-nginx-admission-patch-s5fq7
5665ce1082fa5 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a9f03b34a3cbfbb26d103a14046ab2c5130a80c3d69d526ff8063d2b37b9fd3f 4 minutes ago Exited create 0 7588d6c1b3726 ingress-nginx-admission-create-287m6
d244f32e00679 docker.io/rancher/local-path-provisioner@sha256:73f712e7af12b06720c35ce75217f904f00e4bd96de79f8db1cf160112e667ef 4 minutes ago Running local-path-provisioner 0 99edf650f528e local-path-provisioner-86d989889c-j9wdv
2497007677f8c registry.k8s.io/metrics-server/metrics-server@sha256:78e46b57096ec75e302fbc853e36359555df5c827bb009ecfe66f97474cc2a5a 4 minutes ago Running metrics-server 0 f794499539882 metrics-server-84c5f94fbc-z8rzz
d395437896ee2 c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6 4 minutes ago Running coredns 0 cb969df31de80 coredns-7c65d6cfc9-ksv2k
cbfe74880d2d7 6e38f40d628db3002f5617342c8872c935de530d867d0f709a2fbda1a302a562 4 minutes ago Running storage-provisioner 0 f3272849c7731 storage-provisioner
9a6bfcbfaf469 docker.io/kindest/kindnetd@sha256:3da053f9c42d9123d34d4582cc77041c013e1419204b9ef180f0b3bffa7769e3 4 minutes ago Running kindnet-cni 0 5a7cc27da2525 kindnet-sn2ww
c1be7640a86c8 505d571f5fd56726488d27af0d9a8e02c6db58f5d62ea51dd10d47de7a0c2d38 4 minutes ago Running kube-proxy 0 9f1718b08cd98 kube-proxy-dw2js
93aca58b0473f 2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4 4 minutes ago Running etcd 0 7d663deb48a89 etcd-addons-109663
c2d7f9e7ddfbc 9499c9960544e80a96c223cdc5d3059dd7c2cc37ea20e7138af4a6e415a49173 4 minutes ago Running kube-apiserver 0 52fae8dd5fd08 kube-apiserver-addons-109663
c7d6c76bcfec7 847c7bc1a541865e150af08318f49d02d0e0cff4a0530fd4ffe369e294dd2856 4 minutes ago Running kube-scheduler 0 3bd33f6501e1d kube-scheduler-addons-109663
bb5423f27c7f2 0486b6c53a1b5af26f2ad2fb89a089e04c6baa6369f8545ab0854f9d62b44503 4 minutes ago Running kube-controller-manager 0 3c1599239d305 kube-controller-manager-addons-109663
==> coredns [d395437896ee29deac13e3b40538f4a61f3995dad89fb4205ccef971888d4190] <==
[INFO] 10.244.0.4:37612 - 40060 "AAAA IN registry.kube-system.svc.cluster.local.cluster.local. udp 70 false 512" NXDOMAIN qr,aa,rd 163 0.000093327s
[INFO] 10.244.0.4:45179 - 13675 "A IN registry.kube-system.svc.cluster.local.us-east4-a.c.k8s-minikube.internal. udp 91 false 512" NXDOMAIN qr,rd,ra 91 0.003755375s
[INFO] 10.244.0.4:45179 - 14007 "AAAA IN registry.kube-system.svc.cluster.local.us-east4-a.c.k8s-minikube.internal. udp 91 false 512" NXDOMAIN qr,rd,ra 91 0.00420891s
[INFO] 10.244.0.4:45946 - 63283 "AAAA IN registry.kube-system.svc.cluster.local.c.k8s-minikube.internal. udp 80 false 512" NXDOMAIN qr,rd,ra 80 0.004315672s
[INFO] 10.244.0.4:45946 - 62934 "A IN registry.kube-system.svc.cluster.local.c.k8s-minikube.internal. udp 80 false 512" NXDOMAIN qr,rd,ra 80 0.004618056s
[INFO] 10.244.0.4:44563 - 2352 "A IN registry.kube-system.svc.cluster.local.google.internal. udp 72 false 512" NXDOMAIN qr,rd,ra 72 0.003613643s
[INFO] 10.244.0.4:44563 - 2631 "AAAA IN registry.kube-system.svc.cluster.local.google.internal. udp 72 false 512" NXDOMAIN qr,rd,ra 72 0.00430746s
[INFO] 10.244.0.4:48506 - 5195 "AAAA IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 149 0.00007273s
[INFO] 10.244.0.4:48506 - 5016 "A IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 110 0.000110423s
[INFO] 10.244.0.20:46163 - 36954 "A IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000193657s
[INFO] 10.244.0.20:51642 - 57430 "AAAA IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000254981s
[INFO] 10.244.0.20:44452 - 1364 "A IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000156169s
[INFO] 10.244.0.20:35680 - 23064 "AAAA IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000221411s
[INFO] 10.244.0.20:51507 - 33886 "AAAA IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000135158s
[INFO] 10.244.0.20:38449 - 8264 "A IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000199952s
[INFO] 10.244.0.20:36591 - 32512 "A IN storage.googleapis.com.us-east4-a.c.k8s-minikube.internal. udp 86 false 1232" NXDOMAIN qr,rd,ra 75 0.006706949s
[INFO] 10.244.0.20:43494 - 28794 "AAAA IN storage.googleapis.com.us-east4-a.c.k8s-minikube.internal. udp 86 false 1232" NXDOMAIN qr,rd,ra 75 0.006820448s
[INFO] 10.244.0.20:47839 - 33089 "AAAA IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.004590375s
[INFO] 10.244.0.20:55512 - 37006 "A IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.005359418s
[INFO] 10.244.0.20:44275 - 58247 "AAAA IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.004085356s
[INFO] 10.244.0.20:40163 - 60073 "A IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.005091282s
[INFO] 10.244.0.20:48232 - 52746 "AAAA IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 240 0.000636345s
[INFO] 10.244.0.20:50476 - 59323 "A IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 420 0.000797507s
[INFO] 10.244.0.27:40836 - 2 "AAAA IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 149 0.000231287s
[INFO] 10.244.0.27:50723 - 3 "A IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 110 0.000201791s
==> describe nodes <==
Name: addons-109663
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-109663
kubernetes.io/os=linux
minikube.k8s.io/commit=22da80be3b90f71512d84256b3df4ef76bd13ff8
minikube.k8s.io/name=addons-109663
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_12_16T10_32_58_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-109663
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/crio/crio.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 16 Dec 2024 10:32:55 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-109663
AcquireTime: <unset>
RenewTime: Mon, 16 Dec 2024 10:37:43 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 16 Dec 2024 10:36:01 +0000 Mon, 16 Dec 2024 10:32:54 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 16 Dec 2024 10:36:01 +0000 Mon, 16 Dec 2024 10:32:54 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 16 Dec 2024 10:36:01 +0000 Mon, 16 Dec 2024 10:32:54 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 16 Dec 2024 10:36:01 +0000 Mon, 16 Dec 2024 10:33:22 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-109663
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859312Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859312Ki
pods: 110
System Info:
Machine ID: c878448df26f4703bfd4f4644cd4f6ef
System UUID: 1d94d62c-1455-428d-baf9-9d8a353f13c2
Boot ID: 9fd10bb4-c61e-4d88-b4b5-bae725bc9632
Kernel Version: 5.15.0-1071-gcp
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: cri-o://1.24.6
Kubelet Version: v1.31.2
Kube-Proxy Version: v1.31.2
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (15 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 3m
default hello-world-app-55bf9c44b4-br7qj 0 (0%) 0 (0%) 0 (0%) 0 (0%) 1s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 2m22s
ingress-nginx ingress-nginx-controller-5f85ff4588-5q5qg 100m (1%) 0 (0%) 90Mi (0%) 0 (0%) 4m33s
kube-system coredns-7c65d6cfc9-ksv2k 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 4m40s
kube-system etcd-addons-109663 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 4m45s
kube-system kindnet-sn2ww 100m (1%) 100m (1%) 50Mi (0%) 50Mi (0%) 4m40s
kube-system kube-apiserver-addons-109663 250m (3%) 0 (0%) 0 (0%) 0 (0%) 4m46s
kube-system kube-controller-manager-addons-109663 200m (2%) 0 (0%) 0 (0%) 0 (0%) 4m45s
kube-system kube-ingress-dns-minikube 0 (0%) 0 (0%) 0 (0%) 0 (0%) 4m35s
kube-system kube-proxy-dw2js 0 (0%) 0 (0%) 0 (0%) 0 (0%) 4m40s
kube-system kube-scheduler-addons-109663 100m (1%) 0 (0%) 0 (0%) 0 (0%) 4m45s
kube-system metrics-server-84c5f94fbc-z8rzz 100m (1%) 0 (0%) 200Mi (0%) 0 (0%) 4m34s
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 4m34s
local-path-storage local-path-provisioner-86d989889c-j9wdv 0 (0%) 0 (0%) 0 (0%) 0 (0%) 4m34s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 1050m (13%) 100m (1%)
memory 510Mi (1%) 220Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 4m38s kube-proxy
Normal Starting 4m46s kubelet Starting kubelet.
Warning CgroupV1 4m46s kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeHasSufficientMemory 4m45s kubelet Node addons-109663 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 4m45s kubelet Node addons-109663 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 4m45s kubelet Node addons-109663 status is now: NodeHasSufficientPID
Normal RegisteredNode 4m41s node-controller Node addons-109663 event: Registered Node addons-109663 in Controller
Normal NodeReady 4m21s kubelet Node addons-109663 status is now: NodeReady
==> dmesg <==
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff de be ee 00 db 5d 08 06
[ +0.004678] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 82 d9 73 09 a8 1d 08 06
[ +8.602351] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 12 ec 78 2d 3c ff 08 06
[ +0.000321] IPv4: martian source 10.244.0.3 from 10.244.0.2, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff 52 82 7d e3 e9 86 08 06
[Dec16 09:19] IPv4: martian source 10.244.0.1 from 10.244.0.4, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff d6 d4 08 d7 58 df 08 06
[ +0.000407] IPv4: martian source 10.244.0.4 from 10.244.0.3, on dev eth0
[ +0.000003] ll header: 00000000: ff ff ff ff ff ff 82 d9 73 09 a8 1d 08 06
[Dec16 10:35] IPv4: martian source 10.244.0.22 from 127.0.0.1, on dev eth0
[ +0.000008] ll header: 00000000: de d7 2f 49 bb 5f fa 9e 56 e2 a0 0e 08 00
[ +1.023752] IPv4: martian source 10.244.0.22 from 127.0.0.1, on dev eth0
[ +0.000006] ll header: 00000000: de d7 2f 49 bb 5f fa 9e 56 e2 a0 0e 08 00
[ +2.015839] IPv4: martian source 10.244.0.22 from 127.0.0.1, on dev eth0
[ +0.000018] ll header: 00000000: de d7 2f 49 bb 5f fa 9e 56 e2 a0 0e 08 00
[ +4.095632] IPv4: martian source 10.244.0.22 from 127.0.0.1, on dev eth0
[ +0.000006] ll header: 00000000: de d7 2f 49 bb 5f fa 9e 56 e2 a0 0e 08 00
[ +8.195350] IPv4: martian source 10.244.0.22 from 127.0.0.1, on dev eth0
[ +0.000006] ll header: 00000000: de d7 2f 49 bb 5f fa 9e 56 e2 a0 0e 08 00
[Dec16 10:36] IPv4: martian source 10.244.0.22 from 127.0.0.1, on dev eth0
[ +0.000006] ll header: 00000000: de d7 2f 49 bb 5f fa 9e 56 e2 a0 0e 08 00
[ +33.277339] IPv4: martian source 10.244.0.22 from 127.0.0.1, on dev eth0
[ +0.000012] ll header: 00000000: de d7 2f 49 bb 5f fa 9e 56 e2 a0 0e 08 00
==> etcd [93aca58b0473f5baf584710e9ca182179cce77cab936414e2e85aba16f5ad4b6] <==
{"level":"info","ts":"2024-12-16T10:33:07.672486Z","caller":"traceutil/trace.go:171","msg":"trace[793751114] range","detail":"{range_begin:/registry/serviceaccounts/kube-system/storage-provisioner; range_end:; response_count:0; response_revision:377; }","duration":"295.672195ms","start":"2024-12-16T10:33:07.376804Z","end":"2024-12-16T10:33:07.672476Z","steps":["trace[793751114] 'agreement among raft nodes before linearized reading' (duration: 216.741433ms)","trace[793751114] 'range keys from in-memory index tree' (duration: 78.883366ms)"],"step_count":2}
{"level":"warn","ts":"2024-12-16T10:33:07.672864Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"186.571774ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/ranges/serviceips\" ","response":"range_response_count:1 size:116"}
{"level":"info","ts":"2024-12-16T10:33:07.672892Z","caller":"traceutil/trace.go:171","msg":"trace[1769695805] range","detail":"{range_begin:/registry/ranges/serviceips; range_end:; response_count:1; response_revision:378; }","duration":"186.604327ms","start":"2024-12-16T10:33:07.486278Z","end":"2024-12-16T10:33:07.672883Z","steps":["trace[1769695805] 'agreement among raft nodes before linearized reading' (duration: 186.519799ms)"],"step_count":1}
{"level":"warn","ts":"2024-12-16T10:33:07.994499Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"112.96805ms","expected-duration":"100ms","prefix":"","request":"header:<ID:8128033944884734075 username:\"kube-apiserver-etcd-client\" auth_revision:1 > txn:<compare:<target:MOD key:\"/registry/pods/kube-system/coredns-7c65d6cfc9-d4bw4\" mod_revision:384 > success:<request_delete_range:<key:\"/registry/pods/kube-system/coredns-7c65d6cfc9-d4bw4\" > > failure:<request_range:<key:\"/registry/pods/kube-system/coredns-7c65d6cfc9-d4bw4\" > >>","response":"size:18"}
{"level":"info","ts":"2024-12-16T10:33:08.071909Z","caller":"traceutil/trace.go:171","msg":"trace[99038856] transaction","detail":"{read_only:false; number_of_response:1; response_revision:392; }","duration":"191.844388ms","start":"2024-12-16T10:33:07.880048Z","end":"2024-12-16T10:33:08.071892Z","steps":["trace[99038856] 'compare' (duration: 112.892348ms)"],"step_count":1}
{"level":"info","ts":"2024-12-16T10:33:08.072065Z","caller":"traceutil/trace.go:171","msg":"trace[1239885436] linearizableReadLoop","detail":"{readStateIndex:407; appliedIndex:406; }","duration":"191.650397ms","start":"2024-12-16T10:33:07.880403Z","end":"2024-12-16T10:33:08.072053Z","steps":["trace[1239885436] 'read index received' (duration: 833.119µs)","trace[1239885436] 'applied index is now lower than readState.Index' (duration: 190.816182ms)"],"step_count":2}
{"level":"info","ts":"2024-12-16T10:33:08.072223Z","caller":"traceutil/trace.go:171","msg":"trace[738296694] transaction","detail":"{read_only:false; response_revision:393; number_of_response:1; }","duration":"191.759501ms","start":"2024-12-16T10:33:07.880456Z","end":"2024-12-16T10:33:08.072215Z","steps":["trace[738296694] 'process raft request' (duration: 114.121491ms)"],"step_count":1}
{"level":"info","ts":"2024-12-16T10:33:08.072325Z","caller":"traceutil/trace.go:171","msg":"trace[817934492] transaction","detail":"{read_only:false; response_revision:394; number_of_response:1; }","duration":"191.792502ms","start":"2024-12-16T10:33:07.880522Z","end":"2024-12-16T10:33:08.072315Z","steps":["trace[817934492] 'process raft request' (duration: 114.113915ms)"],"step_count":1}
{"level":"info","ts":"2024-12-16T10:33:08.072437Z","caller":"traceutil/trace.go:171","msg":"trace[960474592] transaction","detail":"{read_only:false; response_revision:395; number_of_response:1; }","duration":"191.735918ms","start":"2024-12-16T10:33:07.880693Z","end":"2024-12-16T10:33:08.072429Z","steps":["trace[960474592] 'process raft request' (duration: 113.973168ms)"],"step_count":1}
{"level":"warn","ts":"2024-12-16T10:33:08.072677Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"192.261081ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/services/specs/kube-system/registry\" ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-12-16T10:33:08.072709Z","caller":"traceutil/trace.go:171","msg":"trace[1404446973] range","detail":"{range_begin:/registry/services/specs/kube-system/registry; range_end:; response_count:0; response_revision:401; }","duration":"192.300009ms","start":"2024-12-16T10:33:07.880400Z","end":"2024-12-16T10:33:08.072700Z","steps":["trace[1404446973] 'agreement among raft nodes before linearized reading' (duration: 192.239715ms)"],"step_count":1}
{"level":"warn","ts":"2024-12-16T10:33:08.072859Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"192.039823ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/ranges/servicenodeports\" ","response":"range_response_count:1 size:260"}
{"level":"info","ts":"2024-12-16T10:33:08.072886Z","caller":"traceutil/trace.go:171","msg":"trace[381039783] range","detail":"{range_begin:/registry/ranges/servicenodeports; range_end:; response_count:1; response_revision:401; }","duration":"192.072803ms","start":"2024-12-16T10:33:07.880806Z","end":"2024-12-16T10:33:08.072879Z","steps":["trace[381039783] 'agreement among raft nodes before linearized reading' (duration: 192.013363ms)"],"step_count":1}
{"level":"warn","ts":"2024-12-16T10:33:08.381071Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"100.234713ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/apiregistration.k8s.io/apiservices/v1beta1.metrics.k8s.io\" ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-12-16T10:33:08.381168Z","caller":"traceutil/trace.go:171","msg":"trace[1952366159] range","detail":"{range_begin:/registry/apiregistration.k8s.io/apiservices/v1beta1.metrics.k8s.io; range_end:; response_count:0; response_revision:425; }","duration":"100.32411ms","start":"2024-12-16T10:33:08.280818Z","end":"2024-12-16T10:33:08.381142Z","steps":["trace[1952366159] 'agreement among raft nodes before linearized reading' (duration: 100.095009ms)"],"step_count":1}
{"level":"warn","ts":"2024-12-16T10:33:08.485917Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"199.755156ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/deployments/kube-system/registry\" ","response":"range_response_count:1 size:3350"}
{"level":"info","ts":"2024-12-16T10:33:08.486052Z","caller":"traceutil/trace.go:171","msg":"trace[816096498] range","detail":"{range_begin:/registry/deployments/kube-system/registry; range_end:; response_count:1; response_revision:426; }","duration":"199.895603ms","start":"2024-12-16T10:33:08.286139Z","end":"2024-12-16T10:33:08.486035Z","steps":["trace[816096498] 'agreement among raft nodes before linearized reading' (duration: 95.906549ms)","trace[816096498] 'range keys from in-memory index tree' (duration: 92.502138ms)"],"step_count":2}
{"level":"info","ts":"2024-12-16T10:33:08.486128Z","caller":"traceutil/trace.go:171","msg":"trace[641752804] transaction","detail":"{read_only:false; response_revision:427; number_of_response:1; }","duration":"103.967584ms","start":"2024-12-16T10:33:08.382149Z","end":"2024-12-16T10:33:08.486117Z","steps":["trace[641752804] 'process raft request' (duration: 91.625966ms)","trace[641752804] 'compare' (duration: 11.954162ms)"],"step_count":2}
{"level":"info","ts":"2024-12-16T10:34:05.693646Z","caller":"traceutil/trace.go:171","msg":"trace[1539685743] transaction","detail":"{read_only:false; response_revision:1111; number_of_response:1; }","duration":"105.470497ms","start":"2024-12-16T10:34:05.588152Z","end":"2024-12-16T10:34:05.693622Z","steps":["trace[1539685743] 'process raft request' (duration: 88.430195ms)","trace[1539685743] 'compare' (duration: 16.935916ms)"],"step_count":2}
{"level":"info","ts":"2024-12-16T10:34:28.616542Z","caller":"traceutil/trace.go:171","msg":"trace[1208903567] linearizableReadLoop","detail":"{readStateIndex:1235; appliedIndex:1234; }","duration":"116.984843ms","start":"2024-12-16T10:34:28.499541Z","end":"2024-12-16T10:34:28.616526Z","steps":["trace[1208903567] 'read index received' (duration: 54.381885ms)","trace[1208903567] 'applied index is now lower than readState.Index' (duration: 62.602481ms)"],"step_count":2}
{"level":"info","ts":"2024-12-16T10:34:28.616682Z","caller":"traceutil/trace.go:171","msg":"trace[608077655] transaction","detail":"{read_only:false; response_revision:1197; number_of_response:1; }","duration":"197.753384ms","start":"2024-12-16T10:34:28.418905Z","end":"2024-12-16T10:34:28.616659Z","steps":["trace[608077655] 'process raft request' (duration: 135.083294ms)","trace[608077655] 'compare' (duration: 62.448167ms)"],"step_count":2}
{"level":"warn","ts":"2024-12-16T10:34:28.616727Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"117.057461ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/leases/kube-system/external-health-monitor-leader-hostpath-csi-k8s-io\" ","response":"range_response_count:1 size:554"}
{"level":"warn","ts":"2024-12-16T10:34:28.616733Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"117.167393ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods/kube-system/metrics-server-84c5f94fbc-z8rzz\" ","response":"range_response_count:1 size:4862"}
{"level":"info","ts":"2024-12-16T10:34:28.616768Z","caller":"traceutil/trace.go:171","msg":"trace[312161016] range","detail":"{range_begin:/registry/leases/kube-system/external-health-monitor-leader-hostpath-csi-k8s-io; range_end:; response_count:1; response_revision:1197; }","duration":"117.108854ms","start":"2024-12-16T10:34:28.499643Z","end":"2024-12-16T10:34:28.616752Z","steps":["trace[312161016] 'agreement among raft nodes before linearized reading' (duration: 116.977084ms)"],"step_count":1}
{"level":"info","ts":"2024-12-16T10:34:28.616774Z","caller":"traceutil/trace.go:171","msg":"trace[1404395514] range","detail":"{range_begin:/registry/pods/kube-system/metrics-server-84c5f94fbc-z8rzz; range_end:; response_count:1; response_revision:1197; }","duration":"117.232126ms","start":"2024-12-16T10:34:28.499532Z","end":"2024-12-16T10:34:28.616764Z","steps":["trace[1404395514] 'agreement among raft nodes before linearized reading' (duration: 117.089497ms)"],"step_count":1}
==> kernel <==
10:37:43 up 3:20, 0 users, load average: 0.14, 30.99, 85.30
Linux addons-109663 5.15.0-1071-gcp #79~20.04.1-Ubuntu SMP Thu Oct 17 21:59:34 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kindnet [9a6bfcbfaf469f95d1c0c8cbed2904943c6b3ed6c03103d1ddd3c1b525a828c9] <==
I1216 10:35:41.673431 1 main.go:301] handling current node
I1216 10:35:51.672575 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:35:51.672625 1 main.go:301] handling current node
I1216 10:36:01.672946 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:36:01.672982 1 main.go:301] handling current node
I1216 10:36:11.673658 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:36:11.673724 1 main.go:301] handling current node
I1216 10:36:21.679547 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:36:21.679588 1 main.go:301] handling current node
I1216 10:36:31.672627 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:36:31.672666 1 main.go:301] handling current node
I1216 10:36:41.679607 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:36:41.679642 1 main.go:301] handling current node
I1216 10:36:51.681302 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:36:51.681337 1 main.go:301] handling current node
I1216 10:37:01.679573 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:37:01.679605 1 main.go:301] handling current node
I1216 10:37:11.673589 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:37:11.673624 1 main.go:301] handling current node
I1216 10:37:21.681414 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:37:21.681454 1 main.go:301] handling current node
I1216 10:37:31.681501 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:37:31.681540 1 main.go:301] handling current node
I1216 10:37:41.681577 1 main.go:297] Handling node with IPs: map[192.168.49.2:{}]
I1216 10:37:41.681608 1 main.go:301] handling current node
==> kube-apiserver [c2d7f9e7ddfbc06209cfd28e0f274033b7c0d8d246840902f5d602801f9c1804] <==
E1216 10:34:32.691074 1 remote_available_controller.go:448] "Unhandled Error" err="v1beta1.metrics.k8s.io failed with: failing or missing response from https://10.104.203.234:443/apis/metrics.k8s.io/v1beta1: Get \"https://10.104.203.234:443/apis/metrics.k8s.io/v1beta1\": dial tcp 10.104.203.234:443: connect: connection refused" logger="UnhandledError"
E1216 10:34:32.692656 1 remote_available_controller.go:448] "Unhandled Error" err="v1beta1.metrics.k8s.io failed with: failing or missing response from https://10.104.203.234:443/apis/metrics.k8s.io/v1beta1: Get \"https://10.104.203.234:443/apis/metrics.k8s.io/v1beta1\": dial tcp 10.104.203.234:443: connect: connection refused" logger="UnhandledError"
I1216 10:34:32.723275 1 handler.go:286] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
E1216 10:34:51.896672 1 conn.go:339] Error on socket receive: read tcp 192.168.49.2:8443->192.168.49.1:38980: use of closed network connection
E1216 10:34:52.055857 1 conn.go:339] Error on socket receive: read tcp 192.168.49.2:8443->192.168.49.1:39006: use of closed network connection
I1216 10:35:01.015105 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.110.13.210"}
I1216 10:35:21.588891 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I1216 10:35:21.753463 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.104.56.95"}
I1216 10:35:23.395373 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W1216 10:35:24.473039 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I1216 10:35:48.692045 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I1216 10:36:01.744832 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I1216 10:36:01.744890 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I1216 10:36:01.758462 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I1216 10:36:01.758509 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I1216 10:36:01.758956 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I1216 10:36:01.759012 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I1216 10:36:01.772864 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I1216 10:36:01.773001 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I1216 10:36:01.783253 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I1216 10:36:01.783298 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W1216 10:36:02.759800 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W1216 10:36:02.784525 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
W1216 10:36:02.880411 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
I1216 10:37:42.590874 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.99.178.29"}
==> kube-controller-manager [bb5423f27c7f2a039f4792caef379046c477b0bd38adbf004fa89dfc343f7b50] <==
E1216 10:36:17.730622 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W1216 10:36:22.386542 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E1216 10:36:22.386593 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W1216 10:36:22.494051 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E1216 10:36:22.494088 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W1216 10:36:32.538411 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E1216 10:36:32.538455 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W1216 10:36:41.463757 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E1216 10:36:41.463802 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W1216 10:36:42.562182 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E1216 10:36:42.562224 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W1216 10:36:42.664122 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E1216 10:36:42.664158 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W1216 10:37:08.405172 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E1216 10:37:08.405218 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W1216 10:37:09.740745 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E1216 10:37:09.740787 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W1216 10:37:10.617185 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E1216 10:37:10.617231 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W1216 10:37:28.877574 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E1216 10:37:28.877637 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I1216 10:37:42.393681 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="12.088603ms"
I1216 10:37:42.398659 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="4.925663ms"
I1216 10:37:42.398743 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="41.342µs"
I1216 10:37:42.404194 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="83.43µs"
==> kube-proxy [c1be7640a86c8f90c664638cd41fcf0e1115f9837b800644bb080433f43935cc] <==
I1216 10:33:04.183324 1 server_linux.go:66] "Using iptables proxy"
I1216 10:33:04.591563 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E1216 10:33:04.678101 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I1216 10:33:05.489706 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I1216 10:33:05.489861 1 server_linux.go:169] "Using iptables Proxier"
I1216 10:33:05.694015 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I1216 10:33:05.694429 1 server.go:483] "Version info" version="v1.31.2"
I1216 10:33:05.694453 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I1216 10:33:05.788634 1 config.go:199] "Starting service config controller"
I1216 10:33:06.171703 1 shared_informer.go:313] Waiting for caches to sync for service config
I1216 10:33:06.171734 1 shared_informer.go:320] Caches are synced for service config
I1216 10:33:05.790467 1 config.go:105] "Starting endpoint slice config controller"
I1216 10:33:06.171782 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I1216 10:33:06.171788 1 shared_informer.go:320] Caches are synced for endpoint slice config
I1216 10:33:05.790422 1 config.go:328] "Starting node config controller"
I1216 10:33:06.171865 1 shared_informer.go:313] Waiting for caches to sync for node config
I1216 10:33:06.171872 1 shared_informer.go:320] Caches are synced for node config
==> kube-scheduler [c7d6c76bcfec710b57c6b1fe3c19335fd182a394babb6ed68c250307fd00cd53] <==
W1216 10:32:55.981327 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E1216 10:32:55.981346 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
E1216 10:32:55.981349 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
E1216 10:32:55.981313 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W1216 10:32:55.981466 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
W1216 10:32:55.981508 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E1216 10:32:55.981539 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W1216 10:32:55.981556 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E1216 10:32:55.981551 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
E1216 10:32:55.981651 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W1216 10:32:55.981595 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E1216 10:32:55.981694 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W1216 10:32:55.981595 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E1216 10:32:55.981721 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W1216 10:32:55.981602 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E1216 10:32:55.981745 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W1216 10:32:56.856381 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E1216 10:32:56.856434 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W1216 10:32:56.883003 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E1216 10:32:56.883044 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W1216 10:32:56.921571 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E1216 10:32:56.921608 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W1216 10:32:56.963959 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E1216 10:32:56.964001 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
I1216 10:32:57.377547 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Dec 16 10:37:38 addons-109663 kubelet[1650]: E1216 10:37:38.231358 1650 eviction_manager.go:257] "Eviction manager: failed to get HasDedicatedImageFs" err="missing image stats: &ImageFsInfoResponse{ImageFilesystems:[]*FilesystemUsage{&FilesystemUsage{Timestamp:1734345458231192433,FsId:&FilesystemIdentifier{Mountpoint:/var/lib/containers/storage/overlay-images,},UsedBytes:&UInt64Value{Value:617956,},InodesUsed:&UInt64Value{Value:236,},},},ContainerFilesystems:[]*FilesystemUsage{},}"
Dec 16 10:37:38 addons-109663 kubelet[1650]: E1216 10:37:38.231392 1650 eviction_manager.go:212] "Eviction manager: failed to synchronize" err="eviction manager: failed to get HasDedicatedImageFs: missing image stats: &ImageFsInfoResponse{ImageFilesystems:[]*FilesystemUsage{&FilesystemUsage{Timestamp:1734345458231192433,FsId:&FilesystemIdentifier{Mountpoint:/var/lib/containers/storage/overlay-images,},UsedBytes:&UInt64Value{Value:617956,},InodesUsed:&UInt64Value{Value:236,},},},ContainerFilesystems:[]*FilesystemUsage{},}"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394676 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="node-driver-registrar"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394725 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="csi-snapshotter"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394736 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="963124d9-8e43-4fb9-a011-05c542d2fb50" containerName="csi-resizer"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394744 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="csi-provisioner"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394754 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="62bd9cad-e4a7-474c-9ce0-bb38412ded35" containerName="volume-snapshot-controller"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394762 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="hostpath"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394772 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="liveness-probe"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394782 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="9089b466-c717-4755-bf51-2740aecfaeb6" containerName="csi-attacher"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394790 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="07292c74-48a5-4558-9412-61806490f959" containerName="task-pv-container"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394799 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="29ea6b74-8543-4d6d-a9f0-8476aaef7f19" containerName="volume-snapshot-controller"
Dec 16 10:37:42 addons-109663 kubelet[1650]: E1216 10:37:42.394807 1650 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="csi-external-health-monitor-controller"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394878 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="csi-external-health-monitor-controller"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394890 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="hostpath"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394900 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="csi-provisioner"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394909 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="963124d9-8e43-4fb9-a011-05c542d2fb50" containerName="csi-resizer"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394916 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="07292c74-48a5-4558-9412-61806490f959" containerName="task-pv-container"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394927 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="9089b466-c717-4755-bf51-2740aecfaeb6" containerName="csi-attacher"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394935 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="csi-snapshotter"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394943 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="29ea6b74-8543-4d6d-a9f0-8476aaef7f19" containerName="volume-snapshot-controller"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394951 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="62bd9cad-e4a7-474c-9ce0-bb38412ded35" containerName="volume-snapshot-controller"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394958 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="node-driver-registrar"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.394967 1650 memory_manager.go:354] "RemoveStaleState removing state" podUID="856ef16b-5b68-404c-8df4-558dc73fe76b" containerName="liveness-probe"
Dec 16 10:37:42 addons-109663 kubelet[1650]: I1216 10:37:42.571791 1650 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8df6\" (UniqueName: \"kubernetes.io/projected/6046c7ab-0532-4ad2-907c-cbe45f15d836-kube-api-access-h8df6\") pod \"hello-world-app-55bf9c44b4-br7qj\" (UID: \"6046c7ab-0532-4ad2-907c-cbe45f15d836\") " pod="default/hello-world-app-55bf9c44b4-br7qj"
==> storage-provisioner [cbfe74880d2d74600d5e828c17a093b09e9242e83f220b8981aab484b98eba00] <==
I1216 10:33:23.101386 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I1216 10:33:23.109387 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I1216 10:33:23.109441 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I1216 10:33:23.119725 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I1216 10:33:23.119895 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-109663_ffc62bba-699e-4bb1-b733-f38ab028cbbd!
I1216 10:33:23.120209 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"172750c2-26af-46b6-a829-2003eae424b5", APIVersion:"v1", ResourceVersion:"892", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-109663_ffc62bba-699e-4bb1-b733-f38ab028cbbd became leader
I1216 10:33:23.272006 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-109663_ffc62bba-699e-4bb1-b733-f38ab028cbbd!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-109663 -n addons-109663
helpers_test.go:261: (dbg) Run: kubectl --context addons-109663 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: ingress-nginx-admission-create-287m6 ingress-nginx-admission-patch-s5fq7
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Ingress]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-109663 describe pod ingress-nginx-admission-create-287m6 ingress-nginx-admission-patch-s5fq7
helpers_test.go:277: (dbg) Non-zero exit: kubectl --context addons-109663 describe pod ingress-nginx-admission-create-287m6 ingress-nginx-admission-patch-s5fq7: exit status 1 (54.138643ms)
** stderr **
Error from server (NotFound): pods "ingress-nginx-admission-create-287m6" not found
Error from server (NotFound): pods "ingress-nginx-admission-patch-s5fq7" not found
** /stderr **
helpers_test.go:279: kubectl --context addons-109663 describe pod ingress-nginx-admission-create-287m6 ingress-nginx-admission-patch-s5fq7: exit status 1
addons_test.go:992: (dbg) Run: out/minikube-linux-amd64 -p addons-109663 addons disable ingress-dns --alsologtostderr -v=1
addons_test.go:992: (dbg) Done: out/minikube-linux-amd64 -p addons-109663 addons disable ingress-dns --alsologtostderr -v=1: (1.496744236s)
addons_test.go:992: (dbg) Run: out/minikube-linux-amd64 -p addons-109663 addons disable ingress --alsologtostderr -v=1
addons_test.go:992: (dbg) Done: out/minikube-linux-amd64 -p addons-109663 addons disable ingress --alsologtostderr -v=1: (7.675734576s)
--- FAIL: TestAddons/parallel/Ingress (152.57s)