=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:328: registry stabilized in 2.139536ms
addons_test.go:330: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-t6q8h" [4342257a-a438-4180-ab98-bcb513d0521a] Running
addons_test.go:330: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.006037787s
addons_test.go:333: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-ljn4r" [0fd59917-12f5-4b55-b4ed-fb31a0b82ca1] Running
addons_test.go:333: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.003798179s
addons_test.go:338: (dbg) Run: kubectl --context addons-305811 delete po -l run=registry-test --now
addons_test.go:343: (dbg) Run: kubectl --context addons-305811 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:343: (dbg) Non-zero exit: kubectl --context addons-305811 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.07642233s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:345: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-305811 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:349: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:357: (dbg) Run: out/minikube-linux-amd64 -p addons-305811 ip
2024/09/27 00:28:02 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:386: (dbg) Run: out/minikube-linux-amd64 -p addons-305811 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-305811
helpers_test.go:235: (dbg) docker inspect addons-305811:
-- stdout --
[
{
"Id": "b0fc4ba1a814db3175b24d76f019687a53c962f2054766b0fe2ff0e720ee2306",
"Created": "2024-09-27T00:15:06.295749015Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 542069,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-27T00:15:06.403841033Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:fba5f082b59effd6acfcb1eed3d3f86a23bd3a65463877f8197a730d49f52a09",
"ResolvConfPath": "/var/lib/docker/containers/b0fc4ba1a814db3175b24d76f019687a53c962f2054766b0fe2ff0e720ee2306/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/b0fc4ba1a814db3175b24d76f019687a53c962f2054766b0fe2ff0e720ee2306/hostname",
"HostsPath": "/var/lib/docker/containers/b0fc4ba1a814db3175b24d76f019687a53c962f2054766b0fe2ff0e720ee2306/hosts",
"LogPath": "/var/lib/docker/containers/b0fc4ba1a814db3175b24d76f019687a53c962f2054766b0fe2ff0e720ee2306/b0fc4ba1a814db3175b24d76f019687a53c962f2054766b0fe2ff0e720ee2306-json.log",
"Name": "/addons-305811",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-305811:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-305811",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/13e03c5de357f6713a33ab6d7d8c892e1ba4f9690bf7c0d644ca9839295c966c-init/diff:/var/lib/docker/overlay2/4dc0ad7fbb75d47f911966c6e6e6fb4593dffb8a20f85d19010fe87cbb979de1/diff",
"MergedDir": "/var/lib/docker/overlay2/13e03c5de357f6713a33ab6d7d8c892e1ba4f9690bf7c0d644ca9839295c966c/merged",
"UpperDir": "/var/lib/docker/overlay2/13e03c5de357f6713a33ab6d7d8c892e1ba4f9690bf7c0d644ca9839295c966c/diff",
"WorkDir": "/var/lib/docker/overlay2/13e03c5de357f6713a33ab6d7d8c892e1ba4f9690bf7c0d644ca9839295c966c/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "volume",
"Name": "addons-305811",
"Source": "/var/lib/docker/volumes/addons-305811/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
},
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
}
],
"Config": {
"Hostname": "addons-305811",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-305811",
"name.minikube.sigs.k8s.io": "addons-305811",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "e6822ae25ddc572ca9d814933bb30630c108946a85ee45ccd12c230d1a12e353",
"SandboxKey": "/var/run/docker/netns/e6822ae25ddc",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33163"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33164"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33167"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33165"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33166"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-305811": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "8bbc91b9dd6f7576471064010bc793765e4619e45aa40707687bf68ed9e49dc8",
"EndpointID": "a85edd26b207167cfd2a8e0c6d4b80c68de439d545f06dfd4a8daeb5d05c495e",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-305811",
"b0fc4ba1a814"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-305811 -n addons-305811
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-305811 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| start | --download-only -p | download-docker-851602 | jenkins | v1.34.0 | 27 Sep 24 00:14 UTC | |
| | download-docker-851602 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p download-docker-851602 | download-docker-851602 | jenkins | v1.34.0 | 27 Sep 24 00:14 UTC | 27 Sep 24 00:14 UTC |
| start | --download-only -p | binary-mirror-456680 | jenkins | v1.34.0 | 27 Sep 24 00:14 UTC | |
| | binary-mirror-456680 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:45515 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-456680 | binary-mirror-456680 | jenkins | v1.34.0 | 27 Sep 24 00:14 UTC | 27 Sep 24 00:14 UTC |
| addons | disable dashboard -p | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:14 UTC | |
| | addons-305811 | | | | | |
| addons | enable dashboard -p | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:14 UTC | |
| | addons-305811 | | | | | |
| start | -p addons-305811 --wait=true | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:14 UTC | 27 Sep 24 00:18 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| addons | addons-305811 addons disable | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:18 UTC | 27 Sep 24 00:18 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | disable cloud-spanner -p | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:26 UTC | 27 Sep 24 00:26 UTC |
| | addons-305811 | | | | | |
| addons | disable inspektor-gadget -p | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:26 UTC | 27 Sep 24 00:27 UTC |
| | addons-305811 | | | | | |
| addons | addons-305811 addons | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:26 UTC | 27 Sep 24 00:26 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | -p addons-305811 | | | | | |
| ssh | addons-305811 ssh curl -s | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| addons | addons-305811 addons disable | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| ip | addons-305811 ip | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| addons | addons-305811 addons disable | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-305811 addons disable | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| ssh | addons-305811 ssh cat | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | /opt/local-path-provisioner/pvc-5e00d761-eca2-4989-8669-0ec284d57222_default_test-pvc/file1 | | | | | |
| addons | addons-305811 addons disable | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | enable headlamp | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | -p addons-305811 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-305811 addons disable | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-305811 addons | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-305811 addons | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:27 UTC | 27 Sep 24 00:27 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-305811 ip | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:28 UTC | 27 Sep 24 00:28 UTC |
| addons | addons-305811 addons disable | addons-305811 | jenkins | v1.34.0 | 27 Sep 24 00:28 UTC | 27 Sep 24 00:28 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/27 00:14:42
Running on machine: ubuntu-20-agent-14
Binary: Built with gc go1.23.0 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0927 00:14:42.874879 541317 out.go:345] Setting OutFile to fd 1 ...
I0927 00:14:42.875039 541317 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0927 00:14:42.875050 541317 out.go:358] Setting ErrFile to fd 2...
I0927 00:14:42.875055 541317 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0927 00:14:42.875239 541317 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19711-533157/.minikube/bin
I0927 00:14:42.875885 541317 out.go:352] Setting JSON to false
I0927 00:14:42.876800 541317 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":7026,"bootTime":1727389057,"procs":170,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1069-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0927 00:14:42.876916 541317 start.go:139] virtualization: kvm guest
I0927 00:14:42.879178 541317 out.go:177] * [addons-305811] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0927 00:14:42.880402 541317 out.go:177] - MINIKUBE_LOCATION=19711
I0927 00:14:42.880467 541317 notify.go:220] Checking for updates...
I0927 00:14:42.882818 541317 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0927 00:14:42.884166 541317 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19711-533157/kubeconfig
I0927 00:14:42.885358 541317 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19711-533157/.minikube
I0927 00:14:42.886804 541317 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0927 00:14:42.888482 541317 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0927 00:14:42.889930 541317 driver.go:394] Setting default libvirt URI to qemu:///system
I0927 00:14:42.914031 541317 docker.go:123] docker version: linux-27.3.1:Docker Engine - Community
I0927 00:14:42.914165 541317 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0927 00:14:42.962647 541317 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-27 00:14:42.953373064 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647931392 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.3.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:[WARNING: br
idge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled] ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.17.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.7] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0927 00:14:42.962764 541317 docker.go:318] overlay module found
I0927 00:14:42.964991 541317 out.go:177] * Using the docker driver based on user configuration
I0927 00:14:42.966455 541317 start.go:297] selected driver: docker
I0927 00:14:42.966477 541317 start.go:901] validating driver "docker" against <nil>
I0927 00:14:42.966491 541317 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0927 00:14:42.967313 541317 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0927 00:14:43.016107 541317 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-27 00:14:43.007187351 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647931392 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.3.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:[WARNING: br
idge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled] ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.17.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.7] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0927 00:14:43.016335 541317 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0927 00:14:43.016675 541317 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0927 00:14:43.018772 541317 out.go:177] * Using Docker driver with root privileges
I0927 00:14:43.020326 541317 cni.go:84] Creating CNI manager for ""
I0927 00:14:43.020408 541317 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0927 00:14:43.020423 541317 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0927 00:14:43.020506 541317 start.go:340] cluster config:
{Name:addons-305811 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-305811 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0927 00:14:43.021975 541317 out.go:177] * Starting "addons-305811" primary control-plane node in "addons-305811" cluster
I0927 00:14:43.023428 541317 cache.go:121] Beginning downloading kic base image for docker with docker
I0927 00:14:43.025082 541317 out.go:177] * Pulling base image v0.0.45-1727108449-19696 ...
I0927 00:14:43.026892 541317 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0927 00:14:43.026950 541317 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19711-533157/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4
I0927 00:14:43.026961 541317 cache.go:56] Caching tarball of preloaded images
I0927 00:14:43.027041 541317 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 in local docker daemon
I0927 00:14:43.027062 541317 preload.go:172] Found /home/jenkins/minikube-integration/19711-533157/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0927 00:14:43.027075 541317 cache.go:59] Finished verifying existence of preloaded tar for v1.31.1 on docker
I0927 00:14:43.027437 541317 profile.go:143] Saving config to /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/config.json ...
I0927 00:14:43.027463 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/config.json: {Name:mk13dc21de4947e0e409307495ba0f692ef49159 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:14:43.043300 541317 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 to local cache
I0927 00:14:43.043436 541317 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 in local cache directory
I0927 00:14:43.043459 541317 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 in local cache directory, skipping pull
I0927 00:14:43.043466 541317 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 exists in cache, skipping pull
I0927 00:14:43.043474 541317 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 as a tarball
I0927 00:14:43.043481 541317 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 from local cache
I0927 00:14:54.833235 541317 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 from cached tarball
I0927 00:14:54.833286 541317 cache.go:194] Successfully downloaded all kic artifacts
I0927 00:14:54.833346 541317 start.go:360] acquireMachinesLock for addons-305811: {Name:mkb8e8e2a6b461c4d6f042b83b27a14d3104a472 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0927 00:14:54.833478 541317 start.go:364] duration metric: took 102.903µs to acquireMachinesLock for "addons-305811"
I0927 00:14:54.833507 541317 start.go:93] Provisioning new machine with config: &{Name:addons-305811 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-305811 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0927 00:14:54.833614 541317 start.go:125] createHost starting for "" (driver="docker")
I0927 00:14:54.835534 541317 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0927 00:14:54.835856 541317 start.go:159] libmachine.API.Create for "addons-305811" (driver="docker")
I0927 00:14:54.835905 541317 client.go:168] LocalClient.Create starting
I0927 00:14:54.836041 541317 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19711-533157/.minikube/certs/ca.pem
I0927 00:14:54.998339 541317 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19711-533157/.minikube/certs/cert.pem
I0927 00:14:55.132697 541317 cli_runner.go:164] Run: docker network inspect addons-305811 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0927 00:14:55.148351 541317 cli_runner.go:211] docker network inspect addons-305811 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0927 00:14:55.148431 541317 network_create.go:284] running [docker network inspect addons-305811] to gather additional debugging logs...
I0927 00:14:55.148448 541317 cli_runner.go:164] Run: docker network inspect addons-305811
W0927 00:14:55.164361 541317 cli_runner.go:211] docker network inspect addons-305811 returned with exit code 1
I0927 00:14:55.164397 541317 network_create.go:287] error running [docker network inspect addons-305811]: docker network inspect addons-305811: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-305811 not found
I0927 00:14:55.164429 541317 network_create.go:289] output of [docker network inspect addons-305811]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-305811 not found
** /stderr **
I0927 00:14:55.164607 541317 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0927 00:14:55.181113 541317 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc0019a6b90}
I0927 00:14:55.181166 541317 network_create.go:124] attempt to create docker network addons-305811 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0927 00:14:55.181217 541317 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-305811 addons-305811
I0927 00:14:55.241203 541317 network_create.go:108] docker network addons-305811 192.168.49.0/24 created
I0927 00:14:55.241241 541317 kic.go:121] calculated static IP "192.168.49.2" for the "addons-305811" container
I0927 00:14:55.241342 541317 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0927 00:14:55.257737 541317 cli_runner.go:164] Run: docker volume create addons-305811 --label name.minikube.sigs.k8s.io=addons-305811 --label created_by.minikube.sigs.k8s.io=true
I0927 00:14:55.274468 541317 oci.go:103] Successfully created a docker volume addons-305811
I0927 00:14:55.274592 541317 cli_runner.go:164] Run: docker run --rm --name addons-305811-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-305811 --entrypoint /usr/bin/test -v addons-305811:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 -d /var/lib
I0927 00:15:02.247169 541317 cli_runner.go:217] Completed: docker run --rm --name addons-305811-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-305811 --entrypoint /usr/bin/test -v addons-305811:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 -d /var/lib: (6.972466624s)
I0927 00:15:02.247203 541317 oci.go:107] Successfully prepared a docker volume addons-305811
I0927 00:15:02.247220 541317 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0927 00:15:02.247243 541317 kic.go:194] Starting extracting preloaded images to volume ...
I0927 00:15:02.247297 541317 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19711-533157/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-305811:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 -I lz4 -xf /preloaded.tar -C /extractDir
I0927 00:15:06.236595 541317 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19711-533157/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-305811:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 -I lz4 -xf /preloaded.tar -C /extractDir: (3.989233332s)
I0927 00:15:06.236631 541317 kic.go:203] duration metric: took 3.989384847s to extract preloaded images to volume ...
W0927 00:15:06.236754 541317 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0927 00:15:06.236862 541317 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0927 00:15:06.281080 541317 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-305811 --name addons-305811 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-305811 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-305811 --network addons-305811 --ip 192.168.49.2 --volume addons-305811:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21
I0927 00:15:06.568660 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Running}}
I0927 00:15:06.586822 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:06.604706 541317 cli_runner.go:164] Run: docker exec addons-305811 stat /var/lib/dpkg/alternatives/iptables
I0927 00:15:06.646504 541317 oci.go:144] the created container "addons-305811" has a running status.
I0927 00:15:06.646557 541317 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa...
I0927 00:15:06.867351 541317 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0927 00:15:06.890410 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:06.913730 541317 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0927 00:15:06.913751 541317 kic_runner.go:114] Args: [docker exec --privileged addons-305811 chown docker:docker /home/docker/.ssh/authorized_keys]
I0927 00:15:06.961885 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:06.979384 541317 machine.go:93] provisionDockerMachine start ...
I0927 00:15:06.979479 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:06.996901 541317 main.go:141] libmachine: Using SSH client type: native
I0927 00:15:06.997145 541317 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33163 <nil> <nil>}
I0927 00:15:06.997163 541317 main.go:141] libmachine: About to run SSH command:
hostname
I0927 00:15:07.187663 541317 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-305811
I0927 00:15:07.187691 541317 ubuntu.go:169] provisioning hostname "addons-305811"
I0927 00:15:07.187749 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:07.207153 541317 main.go:141] libmachine: Using SSH client type: native
I0927 00:15:07.207381 541317 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33163 <nil> <nil>}
I0927 00:15:07.207399 541317 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-305811 && echo "addons-305811" | sudo tee /etc/hostname
I0927 00:15:07.339592 541317 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-305811
I0927 00:15:07.339677 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:07.360474 541317 main.go:141] libmachine: Using SSH client type: native
I0927 00:15:07.360678 541317 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33163 <nil> <nil>}
I0927 00:15:07.360695 541317 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-305811' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-305811/g' /etc/hosts;
else
echo '127.0.1.1 addons-305811' | sudo tee -a /etc/hosts;
fi
fi
I0927 00:15:07.480680 541317 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0927 00:15:07.480721 541317 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19711-533157/.minikube CaCertPath:/home/jenkins/minikube-integration/19711-533157/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19711-533157/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19711-533157/.minikube}
I0927 00:15:07.480752 541317 ubuntu.go:177] setting up certificates
I0927 00:15:07.480768 541317 provision.go:84] configureAuth start
I0927 00:15:07.480831 541317 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-305811
I0927 00:15:07.497713 541317 provision.go:143] copyHostCerts
I0927 00:15:07.497816 541317 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19711-533157/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19711-533157/.minikube/ca.pem (1078 bytes)
I0927 00:15:07.498002 541317 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19711-533157/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19711-533157/.minikube/cert.pem (1123 bytes)
I0927 00:15:07.498090 541317 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19711-533157/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19711-533157/.minikube/key.pem (1675 bytes)
I0927 00:15:07.498181 541317 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19711-533157/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19711-533157/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19711-533157/.minikube/certs/ca-key.pem org=jenkins.addons-305811 san=[127.0.0.1 192.168.49.2 addons-305811 localhost minikube]
I0927 00:15:07.549587 541317 provision.go:177] copyRemoteCerts
I0927 00:15:07.549653 541317 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0927 00:15:07.549701 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:07.566529 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:07.653088 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I0927 00:15:07.675878 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0927 00:15:07.698657 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0927 00:15:07.721818 541317 provision.go:87] duration metric: took 241.032675ms to configureAuth
I0927 00:15:07.721857 541317 ubuntu.go:193] setting minikube options for container-runtime
I0927 00:15:07.722028 541317 config.go:182] Loaded profile config "addons-305811": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0927 00:15:07.722075 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:07.739255 541317 main.go:141] libmachine: Using SSH client type: native
I0927 00:15:07.739488 541317 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33163 <nil> <nil>}
I0927 00:15:07.739509 541317 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0927 00:15:07.852757 541317 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0927 00:15:07.852787 541317 ubuntu.go:71] root file system type: overlay
I0927 00:15:07.852935 541317 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0927 00:15:07.853009 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:07.870666 541317 main.go:141] libmachine: Using SSH client type: native
I0927 00:15:07.870893 541317 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33163 <nil> <nil>}
I0927 00:15:07.870984 541317 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0927 00:15:07.995367 541317 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0927 00:15:07.995444 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:08.013112 541317 main.go:141] libmachine: Using SSH client type: native
I0927 00:15:08.013312 541317 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33163 <nil> <nil>}
I0927 00:15:08.013331 541317 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0927 00:15:08.692705 541317 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-09-20 11:39:29.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-27 00:15:07.990112941 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0927 00:15:08.692747 541317 machine.go:96] duration metric: took 1.713335936s to provisionDockerMachine
I0927 00:15:08.692762 541317 client.go:171] duration metric: took 13.856847778s to LocalClient.Create
I0927 00:15:08.692786 541317 start.go:167] duration metric: took 13.856934956s to libmachine.API.Create "addons-305811"
I0927 00:15:08.692800 541317 start.go:293] postStartSetup for "addons-305811" (driver="docker")
I0927 00:15:08.692814 541317 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0927 00:15:08.692888 541317 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0927 00:15:08.692938 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:08.710050 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:08.797311 541317 ssh_runner.go:195] Run: cat /etc/os-release
I0927 00:15:08.800330 541317 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0927 00:15:08.800362 541317 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0927 00:15:08.800370 541317 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0927 00:15:08.800377 541317 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0927 00:15:08.800390 541317 filesync.go:126] Scanning /home/jenkins/minikube-integration/19711-533157/.minikube/addons for local assets ...
I0927 00:15:08.800451 541317 filesync.go:126] Scanning /home/jenkins/minikube-integration/19711-533157/.minikube/files for local assets ...
I0927 00:15:08.800476 541317 start.go:296] duration metric: took 107.670076ms for postStartSetup
I0927 00:15:08.800743 541317 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-305811
I0927 00:15:08.817059 541317 profile.go:143] Saving config to /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/config.json ...
I0927 00:15:08.817362 541317 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0927 00:15:08.817416 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:08.834425 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:08.917036 541317 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0927 00:15:08.921474 541317 start.go:128] duration metric: took 14.087836953s to createHost
I0927 00:15:08.921504 541317 start.go:83] releasing machines lock for "addons-305811", held for 14.088011829s
I0927 00:15:08.921583 541317 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-305811
I0927 00:15:08.939627 541317 ssh_runner.go:195] Run: cat /version.json
I0927 00:15:08.939692 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:08.939704 541317 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0927 00:15:08.939766 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:08.956987 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:08.957182 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:09.036230 541317 ssh_runner.go:195] Run: systemctl --version
I0927 00:15:09.109174 541317 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0927 00:15:09.113623 541317 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0927 00:15:09.137561 541317 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0927 00:15:09.137656 541317 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0927 00:15:09.164423 541317 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0927 00:15:09.164450 541317 start.go:495] detecting cgroup driver to use...
I0927 00:15:09.164483 541317 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0927 00:15:09.164598 541317 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0927 00:15:09.179488 541317 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0927 00:15:09.188769 541317 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0927 00:15:09.197899 541317 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0927 00:15:09.197968 541317 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0927 00:15:09.207357 541317 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0927 00:15:09.216703 541317 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0927 00:15:09.226398 541317 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0927 00:15:09.236225 541317 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0927 00:15:09.245176 541317 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0927 00:15:09.254280 541317 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0927 00:15:09.263474 541317 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0927 00:15:09.272720 541317 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0927 00:15:09.280412 541317 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0927 00:15:09.288326 541317 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0927 00:15:09.367383 541317 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0927 00:15:09.453885 541317 start.go:495] detecting cgroup driver to use...
I0927 00:15:09.453939 541317 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0927 00:15:09.453994 541317 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0927 00:15:09.465497 541317 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0927 00:15:09.465557 541317 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0927 00:15:09.476506 541317 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0927 00:15:09.491983 541317 ssh_runner.go:195] Run: which cri-dockerd
I0927 00:15:09.495298 541317 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0927 00:15:09.503981 541317 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0927 00:15:09.530251 541317 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0927 00:15:09.628943 541317 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0927 00:15:09.721580 541317 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0927 00:15:09.721729 541317 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0927 00:15:09.739372 541317 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0927 00:15:09.822180 541317 ssh_runner.go:195] Run: sudo systemctl restart docker
I0927 00:15:10.081820 541317 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0927 00:15:10.092750 541317 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0927 00:15:10.103786 541317 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0927 00:15:10.192212 541317 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0927 00:15:10.275471 541317 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0927 00:15:10.351786 541317 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0927 00:15:10.364378 541317 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0927 00:15:10.374760 541317 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0927 00:15:10.447567 541317 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0927 00:15:10.509398 541317 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0927 00:15:10.509476 541317 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0927 00:15:10.513456 541317 start.go:563] Will wait 60s for crictl version
I0927 00:15:10.513534 541317 ssh_runner.go:195] Run: which crictl
I0927 00:15:10.516854 541317 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0927 00:15:10.550364 541317 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.3.1
RuntimeApiVersion: v1
I0927 00:15:10.550433 541317 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0927 00:15:10.575221 541317 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0927 00:15:10.601621 541317 out.go:235] * Preparing Kubernetes v1.31.1 on Docker 27.3.1 ...
I0927 00:15:10.601722 541317 cli_runner.go:164] Run: docker network inspect addons-305811 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0927 00:15:10.618312 541317 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0927 00:15:10.621916 541317 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0927 00:15:10.632420 541317 kubeadm.go:883] updating cluster {Name:addons-305811 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-305811 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0927 00:15:10.632578 541317 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0927 00:15:10.632643 541317 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0927 00:15:10.652338 541317 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0927 00:15:10.652360 541317 docker.go:615] Images already preloaded, skipping extraction
I0927 00:15:10.652410 541317 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0927 00:15:10.671876 541317 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0927 00:15:10.671902 541317 cache_images.go:84] Images are preloaded, skipping loading
I0927 00:15:10.671914 541317 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.1 docker true true} ...
I0927 00:15:10.672017 541317 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-305811 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.1 ClusterName:addons-305811 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0927 00:15:10.672091 541317 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0927 00:15:10.718306 541317 cni.go:84] Creating CNI manager for ""
I0927 00:15:10.718346 541317 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0927 00:15:10.718364 541317 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0927 00:15:10.718392 541317 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-305811 NodeName:addons-305811 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0927 00:15:10.718546 541317 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-305811"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0927 00:15:10.718613 541317 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.1
I0927 00:15:10.727392 541317 binaries.go:44] Found k8s binaries, skipping transfer
I0927 00:15:10.727462 541317 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0927 00:15:10.736145 541317 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0927 00:15:10.753066 541317 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0927 00:15:10.769898 541317 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0927 00:15:10.788142 541317 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0927 00:15:10.791585 541317 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0927 00:15:10.801672 541317 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0927 00:15:10.885732 541317 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0927 00:15:10.899978 541317 certs.go:68] Setting up /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811 for IP: 192.168.49.2
I0927 00:15:10.900000 541317 certs.go:194] generating shared ca certs ...
I0927 00:15:10.900018 541317 certs.go:226] acquiring lock for ca certs: {Name:mkc81b3bb6b708f82be3877f9c1578c3a8a5359c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:10.900151 541317 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19711-533157/.minikube/ca.key
I0927 00:15:10.991114 541317 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19711-533157/.minikube/ca.crt ...
I0927 00:15:10.991148 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/ca.crt: {Name:mk58b747beb33f905f27b904504ff71bf5ab6633 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:10.991349 541317 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19711-533157/.minikube/ca.key ...
I0927 00:15:10.991368 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/ca.key: {Name:mkde7cddebc6b0ee8713d23a6c6d186b75d84913 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:10.991473 541317 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19711-533157/.minikube/proxy-client-ca.key
I0927 00:15:11.135894 541317 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19711-533157/.minikube/proxy-client-ca.crt ...
I0927 00:15:11.135934 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/proxy-client-ca.crt: {Name:mk8f70a6b358dfa5f84d61ce900d3d7de1e32299 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:11.136140 541317 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19711-533157/.minikube/proxy-client-ca.key ...
I0927 00:15:11.136154 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/proxy-client-ca.key: {Name:mk20507882def3c6664e1715f87527afeaf4b28e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:11.136275 541317 certs.go:256] generating profile certs ...
I0927 00:15:11.136345 541317 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/client.key
I0927 00:15:11.136360 541317 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/client.crt with IP's: []
I0927 00:15:11.176358 541317 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/client.crt ...
I0927 00:15:11.176389 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/client.crt: {Name:mkff39f93fab70eaa7f8bf418505d3479a4f47fe Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:11.176588 541317 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/client.key ...
I0927 00:15:11.176602 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/client.key: {Name:mk8c301144e8f38434f17c5f0206d6f241657ccd Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:11.176696 541317 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.key.db14e583
I0927 00:15:11.176717 541317 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.crt.db14e583 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0927 00:15:11.272848 541317 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.crt.db14e583 ...
I0927 00:15:11.272889 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.crt.db14e583: {Name:mk2ca275f254f6f72b3bab500789d2c3340786eb Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:11.273092 541317 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.key.db14e583 ...
I0927 00:15:11.273109 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.key.db14e583: {Name:mk286acc13044fb0c6f747e485836cad32617da6 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:11.273231 541317 certs.go:381] copying /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.crt.db14e583 -> /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.crt
I0927 00:15:11.273332 541317 certs.go:385] copying /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.key.db14e583 -> /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.key
I0927 00:15:11.273402 541317 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/proxy-client.key
I0927 00:15:11.273428 541317 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/proxy-client.crt with IP's: []
I0927 00:15:11.522580 541317 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/proxy-client.crt ...
I0927 00:15:11.522616 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/proxy-client.crt: {Name:mkdb44ebcccf072bcf9a8d7983555e78eece4cc8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:11.522810 541317 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/proxy-client.key ...
I0927 00:15:11.522826 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/proxy-client.key: {Name:mkd0913639c06e56bc005a2c042c388f24c0e0b8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:11.523028 541317 certs.go:484] found cert: /home/jenkins/minikube-integration/19711-533157/.minikube/certs/ca-key.pem (1679 bytes)
I0927 00:15:11.523075 541317 certs.go:484] found cert: /home/jenkins/minikube-integration/19711-533157/.minikube/certs/ca.pem (1078 bytes)
I0927 00:15:11.523115 541317 certs.go:484] found cert: /home/jenkins/minikube-integration/19711-533157/.minikube/certs/cert.pem (1123 bytes)
I0927 00:15:11.523151 541317 certs.go:484] found cert: /home/jenkins/minikube-integration/19711-533157/.minikube/certs/key.pem (1675 bytes)
I0927 00:15:11.523807 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0927 00:15:11.547254 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0927 00:15:11.569424 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0927 00:15:11.592986 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0927 00:15:11.616521 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0927 00:15:11.639097 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0927 00:15:11.662284 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0927 00:15:11.685102 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/profiles/addons-305811/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0927 00:15:11.707890 541317 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19711-533157/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0927 00:15:11.732635 541317 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0927 00:15:11.749780 541317 ssh_runner.go:195] Run: openssl version
I0927 00:15:11.754916 541317 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0927 00:15:11.764172 541317 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0927 00:15:11.767528 541317 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 27 00:15 /usr/share/ca-certificates/minikubeCA.pem
I0927 00:15:11.767595 541317 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0927 00:15:11.773941 541317 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0927 00:15:11.782806 541317 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0927 00:15:11.785965 541317 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0927 00:15:11.786010 541317 kubeadm.go:392] StartCluster: {Name:addons-305811 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-305811 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0927 00:15:11.786122 541317 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0927 00:15:11.803660 541317 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0927 00:15:11.811930 541317 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0927 00:15:11.820192 541317 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0927 00:15:11.820273 541317 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0927 00:15:11.828478 541317 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0927 00:15:11.828499 541317 kubeadm.go:157] found existing configuration files:
I0927 00:15:11.828545 541317 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0927 00:15:11.836617 541317 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0927 00:15:11.836671 541317 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0927 00:15:11.844589 541317 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0927 00:15:11.852470 541317 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0927 00:15:11.852526 541317 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0927 00:15:11.860342 541317 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0927 00:15:11.868650 541317 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0927 00:15:11.868711 541317 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0927 00:15:11.876651 541317 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0927 00:15:11.884346 541317 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0927 00:15:11.884418 541317 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0927 00:15:11.892420 541317 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0927 00:15:11.926539 541317 kubeadm.go:310] W0927 00:15:11.925835 1918 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0927 00:15:11.927040 541317 kubeadm.go:310] W0927 00:15:11.926476 1918 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0927 00:15:11.947882 541317 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1069-gcp\n", err: exit status 1
I0927 00:15:12.002021 541317 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0927 00:15:21.638277 541317 kubeadm.go:310] [init] Using Kubernetes version: v1.31.1
I0927 00:15:21.638332 541317 kubeadm.go:310] [preflight] Running pre-flight checks
I0927 00:15:21.638399 541317 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0927 00:15:21.638446 541317 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1069-gcp[0m
I0927 00:15:21.638483 541317 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0927 00:15:21.638522 541317 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0927 00:15:21.638575 541317 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0927 00:15:21.638621 541317 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0927 00:15:21.638778 541317 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0927 00:15:21.638896 541317 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0927 00:15:21.639023 541317 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0927 00:15:21.639126 541317 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0927 00:15:21.639196 541317 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0927 00:15:21.639270 541317 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0927 00:15:21.639370 541317 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0927 00:15:21.639519 541317 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0927 00:15:21.639647 541317 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0927 00:15:21.639733 541317 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0927 00:15:21.642354 541317 out.go:235] - Generating certificates and keys ...
I0927 00:15:21.642488 541317 kubeadm.go:310] [certs] Using existing ca certificate authority
I0927 00:15:21.642608 541317 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0927 00:15:21.642695 541317 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0927 00:15:21.642771 541317 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0927 00:15:21.642833 541317 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0927 00:15:21.642887 541317 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0927 00:15:21.642930 541317 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0927 00:15:21.643025 541317 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-305811 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0927 00:15:21.643069 541317 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0927 00:15:21.643170 541317 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-305811 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0927 00:15:21.643262 541317 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0927 00:15:21.643360 541317 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0927 00:15:21.643405 541317 kubeadm.go:310] [certs] Generating "sa" key and public key
I0927 00:15:21.643454 541317 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0927 00:15:21.643499 541317 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0927 00:15:21.643566 541317 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0927 00:15:21.643698 541317 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0927 00:15:21.643788 541317 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0927 00:15:21.643837 541317 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0927 00:15:21.643911 541317 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0927 00:15:21.643991 541317 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0927 00:15:21.645278 541317 out.go:235] - Booting up control plane ...
I0927 00:15:21.645378 541317 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0927 00:15:21.645449 541317 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0927 00:15:21.645528 541317 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0927 00:15:21.645618 541317 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0927 00:15:21.645704 541317 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0927 00:15:21.645749 541317 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0927 00:15:21.645857 541317 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0927 00:15:21.645941 541317 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0927 00:15:21.645988 541317 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.001767541s
I0927 00:15:21.646049 541317 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0927 00:15:21.646098 541317 kubeadm.go:310] [api-check] The API server is healthy after 5.001439361s
I0927 00:15:21.646254 541317 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0927 00:15:21.646397 541317 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0927 00:15:21.646467 541317 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0927 00:15:21.646642 541317 kubeadm.go:310] [mark-control-plane] Marking the node addons-305811 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0927 00:15:21.646729 541317 kubeadm.go:310] [bootstrap-token] Using token: 5vycpi.1lutxmqgxzlgar3f
I0927 00:15:21.648298 541317 out.go:235] - Configuring RBAC rules ...
I0927 00:15:21.648440 541317 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0927 00:15:21.648541 541317 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0927 00:15:21.648732 541317 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0927 00:15:21.648912 541317 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0927 00:15:21.649054 541317 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0927 00:15:21.649138 541317 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0927 00:15:21.649288 541317 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0927 00:15:21.649335 541317 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0927 00:15:21.649374 541317 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0927 00:15:21.649380 541317 kubeadm.go:310]
I0927 00:15:21.649471 541317 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0927 00:15:21.649484 541317 kubeadm.go:310]
I0927 00:15:21.649591 541317 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0927 00:15:21.649602 541317 kubeadm.go:310]
I0927 00:15:21.649638 541317 kubeadm.go:310] mkdir -p $HOME/.kube
I0927 00:15:21.649714 541317 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0927 00:15:21.649771 541317 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0927 00:15:21.649778 541317 kubeadm.go:310]
I0927 00:15:21.649837 541317 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0927 00:15:21.649851 541317 kubeadm.go:310]
I0927 00:15:21.649926 541317 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0927 00:15:21.649940 541317 kubeadm.go:310]
I0927 00:15:21.650013 541317 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0927 00:15:21.650103 541317 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0927 00:15:21.650158 541317 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0927 00:15:21.650164 541317 kubeadm.go:310]
I0927 00:15:21.650296 541317 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0927 00:15:21.650443 541317 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0927 00:15:21.650460 541317 kubeadm.go:310]
I0927 00:15:21.650631 541317 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 5vycpi.1lutxmqgxzlgar3f \
I0927 00:15:21.650788 541317 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:ac657e52bff6dca96e59ceffca02c195587919be0fe004213304b18f9fdc454d \
I0927 00:15:21.650826 541317 kubeadm.go:310] --control-plane
I0927 00:15:21.650836 541317 kubeadm.go:310]
I0927 00:15:21.650934 541317 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0927 00:15:21.650952 541317 kubeadm.go:310]
I0927 00:15:21.651030 541317 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 5vycpi.1lutxmqgxzlgar3f \
I0927 00:15:21.651174 541317 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:ac657e52bff6dca96e59ceffca02c195587919be0fe004213304b18f9fdc454d
I0927 00:15:21.651211 541317 cni.go:84] Creating CNI manager for ""
I0927 00:15:21.651251 541317 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0927 00:15:21.652796 541317 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0927 00:15:21.654031 541317 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0927 00:15:21.663297 541317 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0927 00:15:21.681417 541317 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0927 00:15:21.681529 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:21.681529 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-305811 minikube.k8s.io/updated_at=2024_09_27T00_15_21_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=eee16a295c071ed5a0e96cbbc00bcd13b2654625 minikube.k8s.io/name=addons-305811 minikube.k8s.io/primary=true
I0927 00:15:21.759710 541317 ops.go:34] apiserver oom_adj: -16
I0927 00:15:21.759857 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:22.260437 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:22.760043 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:23.260517 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:23.760061 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:24.260702 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:24.760800 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:25.260207 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:25.760385 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:26.260349 541317 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0927 00:15:26.375448 541317 kubeadm.go:1113] duration metric: took 4.693987598s to wait for elevateKubeSystemPrivileges
I0927 00:15:26.375488 541317 kubeadm.go:394] duration metric: took 14.58948217s to StartCluster
I0927 00:15:26.375508 541317 settings.go:142] acquiring lock: {Name:mk8b12532e2edf4c5c47d5e287e09556e5b5a1af Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:26.375611 541317 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19711-533157/kubeconfig
I0927 00:15:26.375984 541317 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19711-533157/kubeconfig: {Name:mk960a2d635e486d4b21ea2b220ee956331e5ba8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0927 00:15:26.376163 541317 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0927 00:15:26.376180 541317 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0927 00:15:26.376311 541317 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0927 00:15:26.376414 541317 config.go:182] Loaded profile config "addons-305811": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0927 00:15:26.376432 541317 addons.go:69] Setting cloud-spanner=true in profile "addons-305811"
I0927 00:15:26.376449 541317 addons.go:234] Setting addon cloud-spanner=true in "addons-305811"
I0927 00:15:26.376417 541317 addons.go:69] Setting yakd=true in profile "addons-305811"
I0927 00:15:26.376487 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.376491 541317 addons.go:234] Setting addon yakd=true in "addons-305811"
I0927 00:15:26.376520 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.377049 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.377071 541317 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-305811"
I0927 00:15:26.377106 541317 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-305811"
I0927 00:15:26.377129 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.377145 541317 addons.go:69] Setting default-storageclass=true in profile "addons-305811"
I0927 00:15:26.377175 541317 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-305811"
I0927 00:15:26.377234 541317 addons.go:69] Setting storage-provisioner=true in profile "addons-305811"
I0927 00:15:26.377259 541317 addons.go:234] Setting addon storage-provisioner=true in "addons-305811"
I0927 00:15:26.377289 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.377483 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.377494 541317 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-305811"
I0927 00:15:26.377515 541317 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-305811"
I0927 00:15:26.377569 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.377728 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.377767 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.377762 541317 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-305811"
I0927 00:15:26.377787 541317 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-305811"
I0927 00:15:26.377056 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.377818 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.377944 541317 addons.go:69] Setting registry=true in profile "addons-305811"
I0927 00:15:26.377964 541317 addons.go:234] Setting addon registry=true in "addons-305811"
I0927 00:15:26.377989 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.378158 541317 addons.go:69] Setting volumesnapshots=true in profile "addons-305811"
I0927 00:15:26.378203 541317 addons.go:234] Setting addon volumesnapshots=true in "addons-305811"
I0927 00:15:26.378245 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.378279 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.378467 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.378788 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.377132 541317 addons.go:69] Setting metrics-server=true in profile "addons-305811"
I0927 00:15:26.378897 541317 addons.go:234] Setting addon metrics-server=true in "addons-305811"
I0927 00:15:26.379092 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.379315 541317 addons.go:69] Setting inspektor-gadget=true in profile "addons-305811"
I0927 00:15:26.379348 541317 addons.go:234] Setting addon inspektor-gadget=true in "addons-305811"
I0927 00:15:26.379382 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.379816 541317 out.go:177] * Verifying Kubernetes components...
I0927 00:15:26.377478 541317 addons.go:69] Setting volcano=true in profile "addons-305811"
I0927 00:15:26.380107 541317 addons.go:234] Setting addon volcano=true in "addons-305811"
I0927 00:15:26.380146 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.380641 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.380785 541317 addons.go:69] Setting ingress=true in profile "addons-305811"
I0927 00:15:26.380823 541317 addons.go:234] Setting addon ingress=true in "addons-305811"
I0927 00:15:26.379823 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.380949 541317 addons.go:69] Setting ingress-dns=true in profile "addons-305811"
I0927 00:15:26.380973 541317 addons.go:234] Setting addon ingress-dns=true in "addons-305811"
I0927 00:15:26.381004 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.381234 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.380903 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.381915 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.383704 541317 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0927 00:15:26.380916 541317 addons.go:69] Setting gcp-auth=true in profile "addons-305811"
I0927 00:15:26.385493 541317 mustload.go:65] Loading cluster: addons-305811
I0927 00:15:26.385713 541317 config.go:182] Loaded profile config "addons-305811": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0927 00:15:26.385978 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.404356 541317 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.24
I0927 00:15:26.406960 541317 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0927 00:15:26.406985 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0927 00:15:26.407042 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.417258 541317 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-305811"
I0927 00:15:26.417299 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.417613 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.418074 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.421795 541317 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0927 00:15:26.423104 541317 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0927 00:15:26.423126 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0927 00:15:26.423189 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.426124 541317 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0927 00:15:26.426155 541317 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0927 00:15:26.427913 541317 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0927 00:15:26.427942 541317 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0927 00:15:26.428006 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.428194 541317 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0927 00:15:26.428326 541317 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0927 00:15:26.428385 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.445068 541317 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0927 00:15:26.445504 541317 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.7
I0927 00:15:26.446346 541317 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0927 00:15:26.447385 541317 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0927 00:15:26.447454 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.449020 541317 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0927 00:15:26.450364 541317 out.go:177] - Using image docker.io/registry:2.8.3
I0927 00:15:26.450477 541317 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0927 00:15:26.451766 541317 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0927 00:15:26.451989 541317 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0927 00:15:26.452005 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0927 00:15:26.452050 541317 addons.go:234] Setting addon default-storageclass=true in "addons-305811"
I0927 00:15:26.452069 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.452091 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.452632 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:26.455109 541317 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0927 00:15:26.458331 541317 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0927 00:15:26.459716 541317 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0927 00:15:26.460825 541317 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0927 00:15:26.466370 541317 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0927 00:15:26.468107 541317 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0927 00:15:26.468139 541317 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0927 00:15:26.468211 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.476252 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.480628 541317 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0927 00:15:26.481855 541317 out.go:177] - Using image docker.io/busybox:stable
I0927 00:15:26.483099 541317 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0927 00:15:26.483119 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0927 00:15:26.483122 541317 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.10.0
I0927 00:15:26.483168 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.485277 541317 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.10.0
I0927 00:15:26.485905 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:26.487565 541317 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.10.0
I0927 00:15:26.489015 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.490552 541317 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0927 00:15:26.490574 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (471825 bytes)
I0927 00:15:26.490627 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.490853 541317 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0927 00:15:26.492636 541317 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0927 00:15:26.492731 541317 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0927 00:15:26.492756 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0927 00:15:26.492815 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.493635 541317 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0927 00:15:26.493661 541317 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0927 00:15:26.493714 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.494681 541317 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0927 00:15:26.496279 541317 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0927 00:15:26.496300 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0927 00:15:26.496360 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.496527 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.502216 541317 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0927 00:15:26.506257 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.506349 541317 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0927 00:15:26.507982 541317 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0927 00:15:26.509663 541317 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0927 00:15:26.509688 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0927 00:15:26.509755 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.509998 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.525223 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.526896 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.528256 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.538665 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.538660 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.543307 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.549531 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.556451 541317 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0927 00:15:26.556479 541317 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0927 00:15:26.556541 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:26.563342 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:26.576434 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
W0927 00:15:26.624040 541317 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0927 00:15:26.624084 541317 retry.go:31] will retry after 358.209753ms: ssh: handshake failed: EOF
I0927 00:15:26.927333 541317 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0927 00:15:26.927463 541317 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0927 00:15:26.928966 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0927 00:15:26.938366 541317 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0927 00:15:26.938469 541317 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0927 00:15:26.945395 541317 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0927 00:15:26.945425 541317 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0927 00:15:27.022733 541317 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0927 00:15:27.022769 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0927 00:15:27.123739 541317 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0927 00:15:27.123776 541317 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0927 00:15:27.130643 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0927 00:15:27.139459 541317 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0927 00:15:27.139546 541317 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0927 00:15:27.141259 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0927 00:15:27.142502 541317 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0927 00:15:27.142554 541317 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0927 00:15:27.224562 541317 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0927 00:15:27.224655 541317 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0927 00:15:27.427298 541317 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0927 00:15:27.427394 541317 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0927 00:15:27.431709 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0927 00:15:27.436468 541317 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0927 00:15:27.436565 541317 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0927 00:15:27.439278 541317 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0927 00:15:27.439369 541317 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0927 00:15:27.441399 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0927 00:15:27.441775 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0927 00:15:27.522633 541317 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0927 00:15:27.522667 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0927 00:15:27.622386 541317 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0927 00:15:27.622499 541317 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0927 00:15:27.629583 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0927 00:15:27.643205 541317 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0927 00:15:27.643291 541317 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0927 00:15:27.736913 541317 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0927 00:15:27.737004 541317 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0927 00:15:27.838159 541317 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0927 00:15:27.838248 541317 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0927 00:15:27.838894 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0927 00:15:27.839863 541317 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0927 00:15:27.839924 541317 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0927 00:15:27.921254 541317 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0927 00:15:27.921310 541317 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0927 00:15:27.934422 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0927 00:15:28.327083 541317 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0927 00:15:28.327110 541317 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0927 00:15:28.425000 541317 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0927 00:15:28.425034 541317 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0927 00:15:28.433076 541317 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0927 00:15:28.433112 541317 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0927 00:15:28.440341 541317 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0927 00:15:28.440423 541317 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0927 00:15:28.534338 541317 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0927 00:15:28.534418 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0927 00:15:28.637479 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0927 00:15:28.640269 541317 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0927 00:15:28.640361 541317 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0927 00:15:29.041482 541317 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0927 00:15:29.041515 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0927 00:15:29.123462 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0927 00:15:29.125714 541317 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0927 00:15:29.125804 541317 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0927 00:15:29.623915 541317 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0927 00:15:29.624012 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0927 00:15:29.630769 541317 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0927 00:15:29.630800 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0927 00:15:29.638836 541317 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (2.71145474s)
I0927 00:15:29.639871 541317 node_ready.go:35] waiting up to 6m0s for node "addons-305811" to be "Ready" ...
I0927 00:15:29.640154 541317 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (2.712668815s)
I0927 00:15:29.640180 541317 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0927 00:15:29.641314 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (2.712319996s)
I0927 00:15:29.643890 541317 node_ready.go:49] node "addons-305811" has status "Ready":"True"
I0927 00:15:29.643917 541317 node_ready.go:38] duration metric: took 4.002354ms for node "addons-305811" to be "Ready" ...
I0927 00:15:29.643928 541317 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0927 00:15:29.725955 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0927 00:15:29.736885 541317 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-6xq2h" in "kube-system" namespace to be "Ready" ...
I0927 00:15:30.030866 541317 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0927 00:15:30.030904 541317 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0927 00:15:30.037873 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0927 00:15:30.141850 541317 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0927 00:15:30.141890 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0927 00:15:30.221945 541317 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-305811" context rescaled to 1 replicas
I0927 00:15:30.823698 541317 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0927 00:15:30.823731 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0927 00:15:31.433043 541317 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0927 00:15:31.433078 541317 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0927 00:15:31.743859 541317 pod_ready.go:103] pod "coredns-7c65d6cfc9-6xq2h" in "kube-system" namespace has status "Ready":"False"
I0927 00:15:31.829124 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0927 00:15:31.922493 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (4.791806475s)
I0927 00:15:31.922899 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (4.781570209s)
I0927 00:15:32.839150 541317 pod_ready.go:93] pod "coredns-7c65d6cfc9-6xq2h" in "kube-system" namespace has status "Ready":"True"
I0927 00:15:32.839246 541317 pod_ready.go:82] duration metric: took 3.10232007s for pod "coredns-7c65d6cfc9-6xq2h" in "kube-system" namespace to be "Ready" ...
I0927 00:15:32.839274 541317 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-xsg85" in "kube-system" namespace to be "Ready" ...
I0927 00:15:32.928304 541317 pod_ready.go:93] pod "coredns-7c65d6cfc9-xsg85" in "kube-system" namespace has status "Ready":"True"
I0927 00:15:32.928634 541317 pod_ready.go:82] duration metric: took 89.316308ms for pod "coredns-7c65d6cfc9-xsg85" in "kube-system" namespace to be "Ready" ...
I0927 00:15:32.928696 541317 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-305811" in "kube-system" namespace to be "Ready" ...
I0927 00:15:32.934708 541317 pod_ready.go:93] pod "etcd-addons-305811" in "kube-system" namespace has status "Ready":"True"
I0927 00:15:32.934737 541317 pod_ready.go:82] duration metric: took 6.017771ms for pod "etcd-addons-305811" in "kube-system" namespace to be "Ready" ...
I0927 00:15:32.934751 541317 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-305811" in "kube-system" namespace to be "Ready" ...
I0927 00:15:32.941401 541317 pod_ready.go:93] pod "kube-apiserver-addons-305811" in "kube-system" namespace has status "Ready":"True"
I0927 00:15:32.941436 541317 pod_ready.go:82] duration metric: took 6.676485ms for pod "kube-apiserver-addons-305811" in "kube-system" namespace to be "Ready" ...
I0927 00:15:32.941450 541317 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-305811" in "kube-system" namespace to be "Ready" ...
I0927 00:15:33.528786 541317 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0927 00:15:33.528968 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:33.554321 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:34.028370 541317 pod_ready.go:93] pod "kube-controller-manager-addons-305811" in "kube-system" namespace has status "Ready":"True"
I0927 00:15:34.028400 541317 pod_ready.go:82] duration metric: took 1.086940969s for pod "kube-controller-manager-addons-305811" in "kube-system" namespace to be "Ready" ...
I0927 00:15:34.028413 541317 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-mndlk" in "kube-system" namespace to be "Ready" ...
I0927 00:15:34.037828 541317 pod_ready.go:93] pod "kube-proxy-mndlk" in "kube-system" namespace has status "Ready":"True"
I0927 00:15:34.037924 541317 pod_ready.go:82] duration metric: took 9.498559ms for pod "kube-proxy-mndlk" in "kube-system" namespace to be "Ready" ...
I0927 00:15:34.037952 541317 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-305811" in "kube-system" namespace to be "Ready" ...
I0927 00:15:34.128572 541317 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0927 00:15:34.346415 541317 pod_ready.go:93] pod "kube-scheduler-addons-305811" in "kube-system" namespace has status "Ready":"True"
I0927 00:15:34.346439 541317 pod_ready.go:82] duration metric: took 308.469097ms for pod "kube-scheduler-addons-305811" in "kube-system" namespace to be "Ready" ...
I0927 00:15:34.346450 541317 pod_ready.go:79] waiting up to 6m0s for pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace to be "Ready" ...
I0927 00:15:34.541430 541317 addons.go:234] Setting addon gcp-auth=true in "addons-305811"
I0927 00:15:34.541566 541317 host.go:66] Checking if "addons-305811" exists ...
I0927 00:15:34.542341 541317 cli_runner.go:164] Run: docker container inspect addons-305811 --format={{.State.Status}}
I0927 00:15:34.561507 541317 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0927 00:15:34.561567 541317 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-305811
I0927 00:15:34.578589 541317 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33163 SSHKeyPath:/home/jenkins/minikube-integration/19711-533157/.minikube/machines/addons-305811/id_rsa Username:docker}
I0927 00:15:36.436075 541317 pod_ready.go:103] pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace has status "Ready":"False"
I0927 00:15:38.938193 541317 pod_ready.go:103] pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace has status "Ready":"False"
I0927 00:15:39.539509 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (12.10770372s)
I0927 00:15:39.539592 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (12.097757454s)
I0927 00:15:39.539641 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (12.098158903s)
I0927 00:15:39.539797 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (11.910099967s)
I0927 00:15:39.539821 541317 addons.go:475] Verifying addon ingress=true in "addons-305811"
I0927 00:15:39.540057 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (11.701104313s)
I0927 00:15:39.540351 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (11.605894527s)
I0927 00:15:39.540371 541317 addons.go:475] Verifying addon registry=true in "addons-305811"
I0927 00:15:39.541010 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (9.815007256s)
W0927 00:15:39.541049 541317 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0927 00:15:39.541069 541317 retry.go:31] will retry after 162.421801ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0927 00:15:39.541154 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (9.503246255s)
I0927 00:15:39.541421 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (10.417408092s)
I0927 00:15:39.541561 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (10.903233359s)
I0927 00:15:39.541604 541317 addons.go:475] Verifying addon metrics-server=true in "addons-305811"
I0927 00:15:39.542851 541317 out.go:177] * Verifying registry addon...
I0927 00:15:39.542944 541317 out.go:177] * Verifying ingress addon...
I0927 00:15:39.543608 541317 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-305811 service yakd-dashboard -n yakd-dashboard
I0927 00:15:39.545630 541317 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0927 00:15:39.546885 541317 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0927 00:15:39.628994 541317 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0927 00:15:39.629079 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:39.631046 541317 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0927 00:15:39.631076 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:39.704190 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0927 00:15:40.125668 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:40.129770 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:40.341081 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (8.511773272s)
I0927 00:15:40.341129 541317 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-305811"
I0927 00:15:40.341232 541317 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.779692297s)
I0927 00:15:40.344281 541317 out.go:177] * Verifying csi-hostpath-driver addon...
I0927 00:15:40.344286 541317 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0927 00:15:40.345558 541317 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0927 00:15:40.346658 541317 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0927 00:15:40.346696 541317 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0927 00:15:40.346715 541317 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0927 00:15:40.430872 541317 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0927 00:15:40.430904 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:40.528969 541317 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0927 00:15:40.528995 541317 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0927 00:15:40.550160 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:40.553106 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:40.624322 541317 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0927 00:15:40.624356 541317 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0927 00:15:40.724415 541317 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0927 00:15:40.926241 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:41.126269 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:41.126477 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:41.353149 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:41.423545 541317 pod_ready.go:103] pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace has status "Ready":"False"
I0927 00:15:41.551072 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:41.651880 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:41.852838 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:42.054545 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:42.055193 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:42.246366 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.542094165s)
I0927 00:15:42.246470 541317 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.522011267s)
I0927 00:15:42.248818 541317 addons.go:475] Verifying addon gcp-auth=true in "addons-305811"
I0927 00:15:42.250641 541317 out.go:177] * Verifying gcp-auth addon...
I0927 00:15:42.252760 541317 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0927 00:15:42.255615 541317 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0927 00:15:42.357819 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:42.549774 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:42.551316 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:42.851321 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:43.050271 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:43.050617 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:43.351212 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:43.550046 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:43.550446 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:43.850387 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:43.851694 541317 pod_ready.go:103] pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace has status "Ready":"False"
I0927 00:15:44.052551 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:44.052971 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:44.350393 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:44.549898 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:44.551626 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:44.851095 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:45.054148 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:45.054604 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:45.359133 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:45.549897 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:45.550355 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:45.851546 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:45.852963 541317 pod_ready.go:103] pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace has status "Ready":"False"
I0927 00:15:46.049238 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:46.050318 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:46.351621 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:46.549538 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:46.550585 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:46.859700 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:47.049591 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:47.050572 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:47.351569 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:47.549911 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:47.550411 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:47.851364 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:48.049723 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:48.050187 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:48.351305 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:48.352889 541317 pod_ready.go:103] pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace has status "Ready":"False"
I0927 00:15:48.550667 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:48.551010 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:48.851616 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:49.048732 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:49.050668 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:49.357676 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:49.549741 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:49.550148 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:49.851348 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:50.049317 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:50.051097 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:50.351993 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:50.353425 541317 pod_ready.go:103] pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace has status "Ready":"False"
I0927 00:15:50.550108 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:50.550758 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:50.851221 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:51.050110 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:51.051266 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:51.350877 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:51.550726 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:51.552114 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:51.851592 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:52.049777 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:52.051136 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:52.351018 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:52.551032 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:52.551274 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:52.851570 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:52.853226 541317 pod_ready.go:103] pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace has status "Ready":"False"
I0927 00:15:53.049810 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:53.050705 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:53.351064 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:53.550570 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:53.551153 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:53.851494 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:54.050151 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:54.050580 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:54.351167 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:54.550676 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:54.550899 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:54.851154 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:55.049884 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:55.050503 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:55.351088 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:55.352059 541317 pod_ready.go:103] pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace has status "Ready":"False"
I0927 00:15:55.550077 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:55.550420 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:55.850779 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:56.051589 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:56.051885 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:56.351256 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:56.352258 541317 pod_ready.go:93] pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace has status "Ready":"True"
I0927 00:15:56.352282 541317 pod_ready.go:82] duration metric: took 22.005824701s for pod "nvidia-device-plugin-daemonset-ffk56" in "kube-system" namespace to be "Ready" ...
I0927 00:15:56.352289 541317 pod_ready.go:39] duration metric: took 26.708349147s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0927 00:15:56.352314 541317 api_server.go:52] waiting for apiserver process to appear ...
I0927 00:15:56.352372 541317 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0927 00:15:56.368849 541317 api_server.go:72] duration metric: took 29.992618305s to wait for apiserver process to appear ...
I0927 00:15:56.368878 541317 api_server.go:88] waiting for apiserver healthz status ...
I0927 00:15:56.368905 541317 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0927 00:15:56.373196 541317 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0927 00:15:56.374129 541317 api_server.go:141] control plane version: v1.31.1
I0927 00:15:56.374155 541317 api_server.go:131] duration metric: took 5.268541ms to wait for apiserver health ...
I0927 00:15:56.374166 541317 system_pods.go:43] waiting for kube-system pods to appear ...
I0927 00:15:56.382229 541317 system_pods.go:59] 17 kube-system pods found
I0927 00:15:56.382289 541317 system_pods.go:61] "coredns-7c65d6cfc9-xsg85" [cc6ecde0-a1f3-4de9-9a9e-ab49546f6d47] Running
I0927 00:15:56.382300 541317 system_pods.go:61] "csi-hostpath-attacher-0" [6aceffc9-ae54-4f9c-9ab6-88c45b6ed352] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0927 00:15:56.382306 541317 system_pods.go:61] "csi-hostpath-resizer-0" [95b0b782-1860-4f43-8264-b11b363f90af] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0927 00:15:56.382314 541317 system_pods.go:61] "csi-hostpathplugin-tnmlb" [f5036697-f552-4527-a157-9db2dee8910d] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0927 00:15:56.382321 541317 system_pods.go:61] "etcd-addons-305811" [401e4e74-6446-43da-b179-e1423e5391c2] Running
I0927 00:15:56.382326 541317 system_pods.go:61] "kube-apiserver-addons-305811" [47eeb04c-2424-43b1-9827-91461fd6a1f6] Running
I0927 00:15:56.382332 541317 system_pods.go:61] "kube-controller-manager-addons-305811" [ff28d494-fc4c-4dcd-8455-44b40749ad8a] Running
I0927 00:15:56.382341 541317 system_pods.go:61] "kube-ingress-dns-minikube" [0e262871-23c0-43f1-b381-9bf4b6bcca66] Running
I0927 00:15:56.382351 541317 system_pods.go:61] "kube-proxy-mndlk" [8e92c803-4d49-4816-9954-31ef0780ebd9] Running
I0927 00:15:56.382357 541317 system_pods.go:61] "kube-scheduler-addons-305811" [9fc20ba2-7c28-4d58-9fe2-0942dc1f387d] Running
I0927 00:15:56.382367 541317 system_pods.go:61] "metrics-server-84c5f94fbc-cczph" [1150d697-b11c-4f84-9466-d293436bf484] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0927 00:15:56.382373 541317 system_pods.go:61] "nvidia-device-plugin-daemonset-ffk56" [b40c20e3-9fd2-41f6-9116-036b5138f4d1] Running
I0927 00:15:56.382379 541317 system_pods.go:61] "registry-66c9cd494c-t6q8h" [4342257a-a438-4180-ab98-bcb513d0521a] Running
I0927 00:15:56.382391 541317 system_pods.go:61] "registry-proxy-ljn4r" [0fd59917-12f5-4b55-b4ed-fb31a0b82ca1] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0927 00:15:56.382400 541317 system_pods.go:61] "snapshot-controller-56fcc65765-5x7m6" [9c5c71ea-d826-4aeb-a56d-116204ffb507] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0927 00:15:56.382410 541317 system_pods.go:61] "snapshot-controller-56fcc65765-qth4p" [82ff93bb-6e90-4ad1-8dc5-995f83e68be2] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0927 00:15:56.382418 541317 system_pods.go:61] "storage-provisioner" [d9c69eee-5c0d-401c-bfc6-c1817614b7a0] Running
I0927 00:15:56.382428 541317 system_pods.go:74] duration metric: took 8.253484ms to wait for pod list to return data ...
I0927 00:15:56.382440 541317 default_sa.go:34] waiting for default service account to be created ...
I0927 00:15:56.385429 541317 default_sa.go:45] found service account: "default"
I0927 00:15:56.385456 541317 default_sa.go:55] duration metric: took 3.005332ms for default service account to be created ...
I0927 00:15:56.385479 541317 system_pods.go:116] waiting for k8s-apps to be running ...
I0927 00:15:56.394501 541317 system_pods.go:86] 17 kube-system pods found
I0927 00:15:56.394533 541317 system_pods.go:89] "coredns-7c65d6cfc9-xsg85" [cc6ecde0-a1f3-4de9-9a9e-ab49546f6d47] Running
I0927 00:15:56.394544 541317 system_pods.go:89] "csi-hostpath-attacher-0" [6aceffc9-ae54-4f9c-9ab6-88c45b6ed352] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0927 00:15:56.394550 541317 system_pods.go:89] "csi-hostpath-resizer-0" [95b0b782-1860-4f43-8264-b11b363f90af] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0927 00:15:56.394558 541317 system_pods.go:89] "csi-hostpathplugin-tnmlb" [f5036697-f552-4527-a157-9db2dee8910d] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0927 00:15:56.394565 541317 system_pods.go:89] "etcd-addons-305811" [401e4e74-6446-43da-b179-e1423e5391c2] Running
I0927 00:15:56.394571 541317 system_pods.go:89] "kube-apiserver-addons-305811" [47eeb04c-2424-43b1-9827-91461fd6a1f6] Running
I0927 00:15:56.394577 541317 system_pods.go:89] "kube-controller-manager-addons-305811" [ff28d494-fc4c-4dcd-8455-44b40749ad8a] Running
I0927 00:15:56.394587 541317 system_pods.go:89] "kube-ingress-dns-minikube" [0e262871-23c0-43f1-b381-9bf4b6bcca66] Running
I0927 00:15:56.394593 541317 system_pods.go:89] "kube-proxy-mndlk" [8e92c803-4d49-4816-9954-31ef0780ebd9] Running
I0927 00:15:56.394608 541317 system_pods.go:89] "kube-scheduler-addons-305811" [9fc20ba2-7c28-4d58-9fe2-0942dc1f387d] Running
I0927 00:15:56.394616 541317 system_pods.go:89] "metrics-server-84c5f94fbc-cczph" [1150d697-b11c-4f84-9466-d293436bf484] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0927 00:15:56.394623 541317 system_pods.go:89] "nvidia-device-plugin-daemonset-ffk56" [b40c20e3-9fd2-41f6-9116-036b5138f4d1] Running
I0927 00:15:56.394629 541317 system_pods.go:89] "registry-66c9cd494c-t6q8h" [4342257a-a438-4180-ab98-bcb513d0521a] Running
I0927 00:15:56.394637 541317 system_pods.go:89] "registry-proxy-ljn4r" [0fd59917-12f5-4b55-b4ed-fb31a0b82ca1] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0927 00:15:56.394652 541317 system_pods.go:89] "snapshot-controller-56fcc65765-5x7m6" [9c5c71ea-d826-4aeb-a56d-116204ffb507] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0927 00:15:56.394660 541317 system_pods.go:89] "snapshot-controller-56fcc65765-qth4p" [82ff93bb-6e90-4ad1-8dc5-995f83e68be2] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0927 00:15:56.394669 541317 system_pods.go:89] "storage-provisioner" [d9c69eee-5c0d-401c-bfc6-c1817614b7a0] Running
I0927 00:15:56.394679 541317 system_pods.go:126] duration metric: took 9.191743ms to wait for k8s-apps to be running ...
I0927 00:15:56.394693 541317 system_svc.go:44] waiting for kubelet service to be running ....
I0927 00:15:56.394748 541317 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0927 00:15:56.426236 541317 system_svc.go:56] duration metric: took 31.532029ms WaitForService to wait for kubelet
I0927 00:15:56.426270 541317 kubeadm.go:582] duration metric: took 30.050043226s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0927 00:15:56.426294 541317 node_conditions.go:102] verifying NodePressure condition ...
I0927 00:15:56.429791 541317 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0927 00:15:56.429823 541317 node_conditions.go:123] node cpu capacity is 8
I0927 00:15:56.429892 541317 node_conditions.go:105] duration metric: took 3.540814ms to run NodePressure ...
I0927 00:15:56.429931 541317 start.go:241] waiting for startup goroutines ...
I0927 00:15:56.550302 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:56.550780 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:56.857841 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:57.049519 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:57.050122 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:57.359050 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:57.550910 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:57.551248 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:57.852100 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:58.049850 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:58.050673 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:58.351791 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:58.551141 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:58.551181 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:58.851499 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:59.050449 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0927 00:15:59.050827 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:59.351645 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:15:59.593065 541317 kapi.go:107] duration metric: took 20.047430433s to wait for kubernetes.io/minikube-addons=registry ...
I0927 00:15:59.593109 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:15:59.852028 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:00.052782 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:00.352254 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:00.551100 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:00.852258 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:01.052336 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:01.426994 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:01.551929 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:01.851884 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:02.051028 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:02.351331 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:02.551347 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:02.851346 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:03.052474 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:03.352656 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:03.551592 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:03.851262 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:04.050419 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:04.351400 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:04.553981 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:04.851592 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:05.052188 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:05.351723 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:05.551617 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:05.850947 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:06.065396 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:06.352151 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:06.552018 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:06.851228 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:07.051641 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:07.358656 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:07.552022 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:07.851793 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:08.051820 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:08.350955 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:08.551314 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:08.851614 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:09.051203 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:09.351446 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:09.622404 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:09.859220 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:10.052171 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:10.352478 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:10.551605 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:10.858251 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:11.051346 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:11.351478 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:11.551946 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:11.852084 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:12.051321 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:12.358449 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:12.551754 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:12.858388 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:13.051484 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:13.358459 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:13.551830 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:13.859803 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:14.053248 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:14.351485 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:14.552010 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:14.866063 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:15.051809 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:15.351754 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:15.551901 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:15.851697 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:16.051445 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:16.352315 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:16.551757 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:16.875664 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:17.052006 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:17.352131 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:17.552334 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:17.851591 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:18.050959 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:18.352439 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:18.551057 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:18.851557 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:19.051292 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:19.351099 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:19.551689 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:19.887583 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:20.052070 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:20.351970 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:20.552266 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:20.852172 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:21.052853 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:21.351964 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:21.552308 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:21.851837 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:22.051822 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:22.436797 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:22.642611 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:22.851304 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:23.051595 541317 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0927 00:16:23.358465 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:23.551651 541317 kapi.go:107] duration metric: took 44.004764673s to wait for app.kubernetes.io/name=ingress-nginx ...
I0927 00:16:23.850553 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:24.360002 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:24.851934 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:25.351347 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:25.851174 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:26.351880 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:26.850956 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:27.351925 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:27.858205 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:28.351837 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0927 00:16:28.850718 541317 kapi.go:107] duration metric: took 48.504057382s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0927 00:17:04.756444 541317 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0927 00:17:04.756469 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:05.256743 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:05.755687 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:06.256685 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:06.756692 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:07.255675 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:07.757062 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:08.255929 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:08.756058 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:09.255888 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:09.756252 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:10.256508 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:10.756537 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:11.256528 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:11.756649 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:12.257501 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:12.756771 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:13.256303 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:13.756181 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:14.256356 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:14.756450 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:15.256873 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:15.755966 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:16.256395 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:16.756481 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:17.256751 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:17.758194 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:18.256854 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:18.756540 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:19.256314 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:19.756765 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:20.256080 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:20.756159 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:21.256466 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:21.756130 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:22.256702 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:22.756282 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:23.256869 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:23.755913 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:24.256332 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:24.756531 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:25.256243 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:25.756049 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:26.256407 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:26.755954 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:27.256065 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:27.756828 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:28.256378 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:28.756257 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:29.256384 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:29.756658 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:30.256350 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:30.756085 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:31.256429 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:31.756593 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:32.256146 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:32.756801 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:33.256249 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:33.756534 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:34.256973 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:34.756055 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:35.256326 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:35.756302 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:36.256869 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:36.755966 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:37.255678 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:37.757306 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:38.256515 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:38.756424 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:39.256462 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:39.756901 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:40.256432 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:40.756256 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:41.256118 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:41.756587 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:42.258352 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:42.757038 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:43.255999 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:43.755984 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:44.256112 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:44.755833 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:45.256022 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:45.756007 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:46.256503 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:46.756619 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:47.256397 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:47.756662 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:48.256410 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:48.756581 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:49.256649 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:49.756491 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:50.256647 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:50.755691 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:51.255607 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:51.756750 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:52.256419 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:52.756914 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:53.255979 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:53.755919 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:54.256451 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:54.756745 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:55.256170 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:55.755710 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:56.255933 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:56.756213 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:57.256336 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:57.756441 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:58.256082 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:58.756066 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:59.256100 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:17:59.756139 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:00.256772 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:00.755427 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:01.255938 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:01.756445 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:02.256569 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:02.756703 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:03.255867 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:03.755661 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:04.257160 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:04.755946 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:05.256308 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:05.756552 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:06.256448 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:06.755847 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:07.255653 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:07.756649 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:08.256074 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:08.756366 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:09.256512 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:09.756627 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:10.257335 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:10.757872 541317 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0927 00:18:11.256159 541317 kapi.go:107] duration metric: took 2m29.00339634s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0927 00:18:11.257818 541317 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-305811 cluster.
I0927 00:18:11.258999 541317 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0927 00:18:11.260308 541317 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0927 00:18:11.261611 541317 out.go:177] * Enabled addons: cloud-spanner, storage-provisioner, storage-provisioner-rancher, volcano, ingress-dns, nvidia-device-plugin, inspektor-gadget, metrics-server, yakd, default-storageclass, volumesnapshots, registry, ingress, csi-hostpath-driver, gcp-auth
I0927 00:18:11.262833 541317 addons.go:510] duration metric: took 2m44.886527354s for enable addons: enabled=[cloud-spanner storage-provisioner storage-provisioner-rancher volcano ingress-dns nvidia-device-plugin inspektor-gadget metrics-server yakd default-storageclass volumesnapshots registry ingress csi-hostpath-driver gcp-auth]
I0927 00:18:11.262883 541317 start.go:246] waiting for cluster config update ...
I0927 00:18:11.262910 541317 start.go:255] writing updated cluster config ...
I0927 00:18:11.263199 541317 ssh_runner.go:195] Run: rm -f paused
I0927 00:18:11.313746 541317 start.go:600] kubectl: 1.31.1, cluster: 1.31.1 (minor skew: 0)
I0927 00:18:11.315484 541317 out.go:177] * Done! kubectl is now configured to use "addons-305811" cluster and "default" namespace by default
==> Docker <==
Sep 27 00:27:43 addons-305811 dockerd[1337]: time="2024-09-27T00:27:43.091879406Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed" spanID=fc0e86eb9fa03c1b traceID=1fae840ba99d2c33206fed3b04fb8455
Sep 27 00:27:43 addons-305811 dockerd[1337]: time="2024-09-27T00:27:43.093528504Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed" spanID=fc0e86eb9fa03c1b traceID=1fae840ba99d2c33206fed3b04fb8455
Sep 27 00:27:43 addons-305811 dockerd[1337]: time="2024-09-27T00:27:43.111113862Z" level=info msg="ignoring event" container=38a77f18d7f1d6c653e9d0c83d865da4a1697bc035afb8611f8fee94eaab9b82 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:49 addons-305811 dockerd[1337]: time="2024-09-27T00:27:49.784119716Z" level=info msg="ignoring event" container=ae73a35cb16613fd6c2dbe3d5798292b051636b73866ed12a7eb200052aaef8a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:49 addons-305811 dockerd[1337]: time="2024-09-27T00:27:49.909908872Z" level=info msg="ignoring event" container=6a2300ae4ab85a062c547a928914220b2ed4bc56fe8f57334539d29fbe31e8ab module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.422326028Z" level=info msg="ignoring event" container=01568ab3cb1d596b2d4442581ac6597d76bde665af788625a85d6d3fdcf859ba module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.521866729Z" level=info msg="ignoring event" container=1d2ea970913b031a8351c2a41f24d9243890ffec74763c3153d25cdb526a3d34 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.522678160Z" level=info msg="ignoring event" container=a94027cc549285c8c203089d2d40f997128bed6bf0b370f349b3f8f5884533e2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.527006279Z" level=info msg="ignoring event" container=88b78ed8122a13d08665d185414bbbc1b9e7ad6df97514e005177ac4b67890a8 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.533572524Z" level=info msg="ignoring event" container=4baf93c4886c2d274c829a14024eb30fc5bbf24c6e794069c2df1f9d6568d7cf module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.535973152Z" level=info msg="ignoring event" container=ff7a3c0846dcd783bf4f36851b90e4223bd43a7a99346dc14995351db2b1ebfc module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.539139272Z" level=info msg="ignoring event" container=87c16eac7466793206cc9d872e9993ac719a92100ca9d16813efed1fdc8fdb23 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.543585865Z" level=info msg="ignoring event" container=1f1f89952c42ae165dd9119090a4ccfba5ff9d2f0f4c0bb30c83da4efca8bc5f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.830039828Z" level=info msg="ignoring event" container=669383b6f594ab6746e57d7e865bc10db69037774504d8013168680f0fafa4f9 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.882572242Z" level=info msg="ignoring event" container=fd672b478c1f55ffedbe24b890e7706d688e4bf4bb59ceb480823ce68665380e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:51 addons-305811 dockerd[1337]: time="2024-09-27T00:27:51.933486575Z" level=info msg="ignoring event" container=896cdad0b0561fe9c1bbe019b3281fb2c5aa2d97989f8ae08c9d74b180841c39 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:57 addons-305811 dockerd[1337]: time="2024-09-27T00:27:57.860898023Z" level=info msg="ignoring event" container=395c83bd8477d8a5c328205a92c39ae655ee52ae44870b35eb49859d0bdae4c6 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:57 addons-305811 dockerd[1337]: time="2024-09-27T00:27:57.861194038Z" level=info msg="ignoring event" container=33d3a52f94f9e6118a0621443bd1afb78a31eb5213ac2c2500cf78329a547aca module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:58 addons-305811 dockerd[1337]: time="2024-09-27T00:27:58.049465306Z" level=info msg="ignoring event" container=1074a2e6c4c19f1dbe14c176114ef4004abca8b77cfa638e7683fd09beee0de0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:27:58 addons-305811 dockerd[1337]: time="2024-09-27T00:27:58.076089259Z" level=info msg="ignoring event" container=5e484d96b85f41f4ce46daaa2ab590bae95a0b128b6f216da843f04c8463f7ff module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:28:02 addons-305811 dockerd[1337]: time="2024-09-27T00:28:02.491582491Z" level=info msg="ignoring event" container=d73a19eb9dd98a5fef9046caf899c7aa43323554f8b22c0c9d5e1932799d3a42 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:28:02 addons-305811 dockerd[1337]: time="2024-09-27T00:28:02.968309246Z" level=info msg="ignoring event" container=0e1265bac22311a4cfc95dd200420ac264d53f97791ef81154582b8dac4eaf23 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:28:03 addons-305811 dockerd[1337]: time="2024-09-27T00:28:03.034442235Z" level=info msg="ignoring event" container=7ac1f7270abb14ba062978ccce74cfce81c4783b19231e4e36f8016eae0cf557 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:28:03 addons-305811 dockerd[1337]: time="2024-09-27T00:28:03.172336932Z" level=info msg="ignoring event" container=d46051a4d490316f98e939bd9e984bd6866477d7a1277e6629fbfb3f0da69d6f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 27 00:28:03 addons-305811 dockerd[1337]: time="2024-09-27T00:28:03.233982021Z" level=info msg="ignoring event" container=1c81b50f85f67369346e887aeaf20c4f860ad68c764a7b374bba52420ce30aef module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
79ef6861f328b ghcr.io/headlamp-k8s/headlamp@sha256:8825bb13459c64dcf9503d836b94b49c97dc831aff7c325a6eed68961388cf9c 46 seconds ago Running headlamp 0 c037500368495 headlamp-7b5c95b59d-s946n
5f904eb6803dc kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 54 seconds ago Running hello-world-app 0 c0f45c25bd227 hello-world-app-55bf9c44b4-dchgd
7c9f0ba1fe7fa nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf About a minute ago Running nginx 0 8457f14f2d781 nginx
d6d4c049b150a gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 cbf96edd347d3 gcp-auth-89d5ffd79-64cjx
5b2808a0c4715 6e38f40d628db 12 minutes ago Running storage-provisioner 0 f982b4331d592 storage-provisioner
18e57aef5df7f c69fa2e9cbf5f 12 minutes ago Running coredns 0 d6b2e41827736 coredns-7c65d6cfc9-xsg85
4d267a6638bc2 60c005f310ff3 12 minutes ago Running kube-proxy 0 a199f77fe3b5f kube-proxy-mndlk
a0ef58e2e4518 2e96e5913fc06 12 minutes ago Running etcd 0 18d0e7f609bc5 etcd-addons-305811
08ecd2eaa1900 9aa1fad941575 12 minutes ago Running kube-scheduler 0 49b345210a871 kube-scheduler-addons-305811
df3206f160abf 6bab7719df100 12 minutes ago Running kube-apiserver 0 8ad8f60cef0e8 kube-apiserver-addons-305811
d86b3c85aec0a 175ffd71cce3d 12 minutes ago Running kube-controller-manager 0 4f556efc0dabd kube-controller-manager-addons-305811
==> coredns [18e57aef5df7] <==
.:53
[INFO] plugin/reload: Running configuration SHA512 = 591cf328cccc12bc490481273e738df59329c62c0b729d94e8b61db9961c2fa5f046dd37f1cf888b953814040d180f52594972691cd6ff41be96639138a43908
CoreDNS-1.11.3
linux/amd64, go1.21.11, a6338e9
[INFO] Reloading
[INFO] plugin/reload: Running configuration SHA512 = 05e3eaddc414b2d71a69b2e2bc6f2681fc1f4d04bcdd3acc1a41457bb7db518208b95ddfc4c9fffedc59c25a8faf458be1af4915a4a3c0d6777cb7a346bc5d86
[INFO] Reloading complete
[INFO] 127.0.0.1:45991 - 55774 "HINFO IN 8122863217496893449.7465899274160903413. udp 57 false 512" NXDOMAIN qr,rd,ra 57 0.009405739s
[INFO] 10.244.0.25:42425 - 20975 "AAAA IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000334898s
[INFO] 10.244.0.25:53820 - 27437 "A IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000477989s
[INFO] 10.244.0.25:54476 - 32860 "A IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000197555s
[INFO] 10.244.0.25:55192 - 49231 "AAAA IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000274559s
[INFO] 10.244.0.25:58333 - 19851 "AAAA IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000146174s
[INFO] 10.244.0.25:55177 - 37399 "A IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000123313s
[INFO] 10.244.0.25:50679 - 42971 "A IN storage.googleapis.com.us-east4-a.c.k8s-minikube.internal. udp 86 false 1232" NXDOMAIN qr,rd,ra 75 0.008207578s
[INFO] 10.244.0.25:48242 - 5181 "AAAA IN storage.googleapis.com.us-east4-a.c.k8s-minikube.internal. udp 86 false 1232" NXDOMAIN qr,rd,ra 75 0.009986008s
[INFO] 10.244.0.25:49871 - 57474 "A IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.007922835s
[INFO] 10.244.0.25:46356 - 6015 "AAAA IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.008171777s
[INFO] 10.244.0.25:60736 - 35943 "A IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.005392845s
[INFO] 10.244.0.25:50860 - 18804 "AAAA IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.006990306s
[INFO] 10.244.0.25:36579 - 58761 "AAAA IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 240 0.000810325s
[INFO] 10.244.0.25:57633 - 26607 "A IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 458 0.000947794s
==> describe nodes <==
Name: addons-305811
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-305811
kubernetes.io/os=linux
minikube.k8s.io/commit=eee16a295c071ed5a0e96cbbc00bcd13b2654625
minikube.k8s.io/name=addons-305811
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_27T00_15_21_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-305811
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Fri, 27 Sep 2024 00:15:18 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-305811
AcquireTime: <unset>
RenewTime: Fri, 27 Sep 2024 00:27:57 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Fri, 27 Sep 2024 00:27:25 +0000 Fri, 27 Sep 2024 00:15:16 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Fri, 27 Sep 2024 00:27:25 +0000 Fri, 27 Sep 2024 00:15:16 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Fri, 27 Sep 2024 00:27:25 +0000 Fri, 27 Sep 2024 00:15:16 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Fri, 27 Sep 2024 00:27:25 +0000 Fri, 27 Sep 2024 00:15:18 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-305811
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859308Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859308Ki
pods: 110
System Info:
Machine ID: b50c9db7800a4cfcb45c554586efdf8d
System UUID: 687f7c57-8fda-4353-86ac-f1aff31ee784
Boot ID: 3a8dbeac-cc10-412c-b9bd-194fdac9ca10
Kernel Version: 5.15.0-1069-gcp
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.3.1
Kubelet Version: v1.31.1
Kube-Proxy Version: v1.31.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (12 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m15s
default hello-world-app-55bf9c44b4-dchgd 0 (0%) 0 (0%) 0 (0%) 0 (0%) 57s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 66s
gcp-auth gcp-auth-89d5ffd79-64cjx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
headlamp headlamp-7b5c95b59d-s946n 0 (0%) 0 (0%) 0 (0%) 0 (0%) 50s
kube-system coredns-7c65d6cfc9-xsg85 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-305811 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-305811 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-305811 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-mndlk 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-305811 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-305811 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-305811 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-305811 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-305811 event: Registered Node addons-305811 in Controller
==> dmesg <==
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 5a c7 4d ae 54 7f 08 06
[ +1.304624] IPv4: martian source 10.244.0.1 from 10.244.0.16, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 1e 49 c1 d5 4d d2 08 06
[ +1.531711] IPv4: martian source 10.244.0.1 from 10.244.0.17, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 8e 1d 3a 62 73 c9 08 06
[ +8.578090] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff e6 08 e7 57 e7 da 08 06
[ +2.205296] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 4a 8d f1 5f 08 95 08 06
[ +0.271135] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff ba 96 82 9b 58 f6 08 06
[ +0.679861] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 2e 05 2f 7f 47 33 08 06
[ +17.534447] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 42 39 13 4e ef 37 08 06
[Sep27 00:17] IPv4: martian source 10.244.0.1 from 10.244.0.23, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 76 32 e2 e4 24 da 08 06
[ +0.103342] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 52 fc 2e 5a cf 2f 08 06
[Sep27 00:18] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 9e c6 a6 f0 ee 6e 08 06
[ +0.000519] IPv4: martian source 10.244.0.25 from 10.244.0.2, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff f6 d9 d1 1c 77 ec 08 06
[Sep27 00:27] IPv4: martian source 10.244.0.28 from 10.244.0.21, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff e6 08 e7 57 e7 da 08 06
==> etcd [a0ef58e2e451] <==
{"level":"info","ts":"2024-09-27T00:15:16.849050Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-27T00:15:16.849186Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-27T00:15:16.849223Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-27T00:15:16.849303Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-27T00:15:16.849564Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-27T00:15:16.850190Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-09-27T00:15:16.850789Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-27T00:16:22.242841Z","caller":"traceutil/trace.go:171","msg":"trace[782403444] transaction","detail":"{read_only:false; response_revision:1161; number_of_response:1; }","duration":"113.073459ms","start":"2024-09-27T00:16:22.129750Z","end":"2024-09-27T00:16:22.242824Z","steps":["trace[782403444] 'process raft request' (duration: 112.954601ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-27T00:17:49.066019Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"143.842957ms","expected-duration":"100ms","prefix":"read-only range ","request":"limit:1 serializable:true keys_only:true ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-27T00:17:49.066134Z","caller":"traceutil/trace.go:171","msg":"trace[1285405886] range","detail":"{range_begin:; range_end:; response_count:0; response_revision:1413; }","duration":"143.977287ms","start":"2024-09-27T00:17:48.922137Z","end":"2024-09-27T00:17:49.066114Z","steps":["trace[1285405886] 'range keys from in-memory index tree' (duration: 143.828491ms)"],"step_count":1}
{"level":"info","ts":"2024-09-27T00:17:49.066146Z","caller":"traceutil/trace.go:171","msg":"trace[1916399814] transaction","detail":"{read_only:false; response_revision:1414; number_of_response:1; }","duration":"145.366275ms","start":"2024-09-27T00:17:48.920762Z","end":"2024-09-27T00:17:49.066128Z","steps":["trace[1916399814] 'process raft request' (duration: 83.553309ms)","trace[1916399814] 'compare' (duration: 61.656442ms)"],"step_count":2}
{"level":"info","ts":"2024-09-27T00:17:49.066247Z","caller":"traceutil/trace.go:171","msg":"trace[1825703020] transaction","detail":"{read_only:false; response_revision:1415; number_of_response:1; }","duration":"140.265902ms","start":"2024-09-27T00:17:48.925969Z","end":"2024-09-27T00:17:49.066235Z","steps":["trace[1825703020] 'process raft request' (duration: 140.110458ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-27T00:17:49.066312Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"140.277937ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/health\" ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-27T00:17:49.066253Z","caller":"traceutil/trace.go:171","msg":"trace[1134986830] linearizableReadLoop","detail":"{readStateIndex:1458; appliedIndex:1456; }","duration":"140.20169ms","start":"2024-09-27T00:17:48.926026Z","end":"2024-09-27T00:17:49.066228Z","steps":["trace[1134986830] 'read index received' (duration: 78.368294ms)","trace[1134986830] 'applied index is now lower than readState.Index' (duration: 61.83098ms)"],"step_count":2}
{"level":"info","ts":"2024-09-27T00:17:49.066338Z","caller":"traceutil/trace.go:171","msg":"trace[1257221601] range","detail":"{range_begin:/registry/health; range_end:; response_count:0; response_revision:1415; }","duration":"140.313792ms","start":"2024-09-27T00:17:48.926017Z","end":"2024-09-27T00:17:49.066331Z","steps":["trace[1257221601] 'agreement among raft nodes before linearized reading' (duration: 140.257474ms)"],"step_count":1}
{"level":"info","ts":"2024-09-27T00:18:32.024413Z","caller":"traceutil/trace.go:171","msg":"trace[918661633] linearizableReadLoop","detail":"{readStateIndex:1596; appliedIndex:1595; }","duration":"132.376598ms","start":"2024-09-27T00:18:31.892010Z","end":"2024-09-27T00:18:32.024387Z","steps":["trace[918661633] 'read index received' (duration: 73.647393ms)","trace[918661633] 'applied index is now lower than readState.Index' (duration: 58.728308ms)"],"step_count":2}
{"level":"info","ts":"2024-09-27T00:18:32.024425Z","caller":"traceutil/trace.go:171","msg":"trace[2130996785] transaction","detail":"{read_only:false; response_revision:1539; number_of_response:1; }","duration":"165.204331ms","start":"2024-09-27T00:18:31.859197Z","end":"2024-09-27T00:18:32.024402Z","steps":["trace[2130996785] 'process raft request' (duration: 106.452332ms)","trace[2130996785] 'compare' (duration: 58.611531ms)"],"step_count":2}
{"level":"warn","ts":"2024-09-27T00:18:32.024573Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"132.144046ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/health\" ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-27T00:18:32.024600Z","caller":"traceutil/trace.go:171","msg":"trace[181380870] range","detail":"{range_begin:/registry/health; range_end:; response_count:0; response_revision:1539; }","duration":"132.170333ms","start":"2024-09-27T00:18:31.892421Z","end":"2024-09-27T00:18:32.024591Z","steps":["trace[181380870] 'agreement among raft nodes before linearized reading' (duration: 132.124507ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-27T00:18:32.024512Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"132.482353ms","expected-duration":"100ms","prefix":"read-only range ","request":"limit:1 keys_only:true ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-27T00:18:32.024642Z","caller":"traceutil/trace.go:171","msg":"trace[1899664917] range","detail":"{range_begin:; range_end:; response_count:0; response_revision:1539; }","duration":"132.625048ms","start":"2024-09-27T00:18:31.892000Z","end":"2024-09-27T00:18:32.024626Z","steps":["trace[1899664917] 'agreement among raft nodes before linearized reading' (duration: 132.46346ms)"],"step_count":1}
{"level":"info","ts":"2024-09-27T00:25:16.867104Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1855}
{"level":"info","ts":"2024-09-27T00:25:16.891984Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1855,"took":"24.300696ms","hash":1653756623,"current-db-size-bytes":8945664,"current-db-size":"8.9 MB","current-db-size-in-use-bytes":4960256,"current-db-size-in-use":"5.0 MB"}
{"level":"info","ts":"2024-09-27T00:25:16.892038Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":1653756623,"revision":1855,"compact-revision":-1}
{"level":"info","ts":"2024-09-27T00:27:17.802431Z","caller":"traceutil/trace.go:171","msg":"trace[1690332082] transaction","detail":"{read_only:false; response_revision:2773; number_of_response:1; }","duration":"121.96025ms","start":"2024-09-27T00:27:17.680450Z","end":"2024-09-27T00:27:17.802410Z","steps":["trace[1690332082] 'process raft request' (duration: 60.327824ms)","trace[1690332082] 'compare' (duration: 61.507791ms)"],"step_count":2}
==> gcp-auth [d6d4c049b150] <==
2024/09/27 00:18:49 Ready to write response ...
2024/09/27 00:18:50 Ready to marshal response ...
2024/09/27 00:18:50 Ready to write response ...
2024/09/27 00:26:58 Ready to marshal response ...
2024/09/27 00:26:58 Ready to write response ...
2024/09/27 00:27:02 Ready to marshal response ...
2024/09/27 00:27:02 Ready to write response ...
2024/09/27 00:27:04 Ready to marshal response ...
2024/09/27 00:27:04 Ready to write response ...
2024/09/27 00:27:04 Ready to marshal response ...
2024/09/27 00:27:04 Ready to write response ...
2024/09/27 00:27:07 Ready to marshal response ...
2024/09/27 00:27:07 Ready to write response ...
2024/09/27 00:27:12 Ready to marshal response ...
2024/09/27 00:27:12 Ready to write response ...
2024/09/27 00:27:14 Ready to marshal response ...
2024/09/27 00:27:14 Ready to write response ...
2024/09/27 00:27:14 Ready to marshal response ...
2024/09/27 00:27:14 Ready to write response ...
2024/09/27 00:27:14 Ready to marshal response ...
2024/09/27 00:27:14 Ready to write response ...
2024/09/27 00:27:19 Ready to marshal response ...
2024/09/27 00:27:19 Ready to write response ...
2024/09/27 00:27:41 Ready to marshal response ...
2024/09/27 00:27:41 Ready to write response ...
==> kernel <==
00:28:04 up 2:10, 0 users, load average: 0.66, 0.39, 1.27
Linux addons-305811 5.15.0-1069-gcp #77~20.04.1-Ubuntu SMP Sun Sep 1 19:39:16 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [df3206f160ab] <==
W0927 00:18:41.145606 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0927 00:18:41.527129 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0927 00:18:41.847525 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0927 00:26:57.741395 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
I0927 00:26:58.233953 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0927 00:26:58.449863 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.100.89.76"}
W0927 00:26:59.043444 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0927 00:27:02.381418 1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
I0927 00:27:08.031533 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.106.228.236"}
I0927 00:27:14.419887 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.97.123.21"}
I0927 00:27:26.295849 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
E0927 00:27:28.407430 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
I0927 00:27:57.613553 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0927 00:27:57.613607 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0927 00:27:57.626802 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0927 00:27:57.626852 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0927 00:27:57.627419 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0927 00:27:57.627454 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0927 00:27:57.643220 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0927 00:27:57.643581 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0927 00:27:57.737312 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0927 00:27:57.737364 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0927 00:27:58.627736 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W0927 00:27:58.738319 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
W0927 00:27:58.747137 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
==> kube-controller-manager [d86b3c85aec0] <==
W0927 00:27:41.008394 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0927 00:27:41.008441 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0927 00:27:51.291744 1 stateful_set.go:466] "StatefulSet has been deleted" logger="statefulset-controller" key="kube-system/csi-hostpath-attacher"
I0927 00:27:51.343595 1 stateful_set.go:466] "StatefulSet has been deleted" logger="statefulset-controller" key="kube-system/csi-hostpath-resizer"
I0927 00:27:51.739279 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-305811"
I0927 00:27:57.825571 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/snapshot-controller-56fcc65765" duration="11.362µs"
E0927 00:27:58.629200 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
E0927 00:27:58.739669 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
E0927 00:27:58.748429 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0927 00:27:59.760063 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0927 00:27:59.760112 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0927 00:28:00.086639 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0927 00:28:00.086681 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0927 00:28:00.170027 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0927 00:28:00.170074 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0927 00:28:00.530911 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="local-path-storage"
W0927 00:28:01.532514 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0927 00:28:01.532558 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0927 00:28:02.735053 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0927 00:28:02.735099 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0927 00:28:02.933237 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="9.518µs"
W0927 00:28:03.267504 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0927 00:28:03.267549 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0927 00:28:04.066983 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0927 00:28:04.067050 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
==> kube-proxy [4d267a6638bc] <==
I0927 00:15:26.450696 1 server_linux.go:66] "Using iptables proxy"
I0927 00:15:26.738989 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0927 00:15:26.739074 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0927 00:15:27.029176 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0927 00:15:27.029293 1 server_linux.go:169] "Using iptables Proxier"
I0927 00:15:27.039446 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0927 00:15:27.039900 1 server.go:483] "Version info" version="v1.31.1"
I0927 00:15:27.039925 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0927 00:15:27.045414 1 config.go:199] "Starting service config controller"
I0927 00:15:27.045484 1 shared_informer.go:313] Waiting for caches to sync for service config
I0927 00:15:27.045519 1 config.go:105] "Starting endpoint slice config controller"
I0927 00:15:27.045523 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0927 00:15:27.045565 1 config.go:328] "Starting node config controller"
I0927 00:15:27.045585 1 shared_informer.go:313] Waiting for caches to sync for node config
I0927 00:15:27.146705 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0927 00:15:27.146778 1 shared_informer.go:320] Caches are synced for service config
I0927 00:15:27.227839 1 shared_informer.go:320] Caches are synced for node config
==> kube-scheduler [08ecd2eaa190] <==
W0927 00:15:18.545104 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0927 00:15:18.545117 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0927 00:15:18.545182 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0927 00:15:18.545208 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0927 00:15:18.545280 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0927 00:15:18.545310 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0927 00:15:18.545345 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0927 00:15:18.545382 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0927 00:15:18.545559 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0927 00:15:18.545585 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0927 00:15:19.410291 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0927 00:15:19.410333 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0927 00:15:19.419761 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0927 00:15:19.419811 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0927 00:15:19.428290 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0927 00:15:19.428335 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0927 00:15:19.491913 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0927 00:15:19.491975 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0927 00:15:19.529342 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0927 00:15:19.529388 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0927 00:15:19.607677 1 reflector.go:561] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0927 00:15:19.607726 1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
W0927 00:15:19.617143 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0927 00:15:19.617184 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
I0927 00:15:22.542920 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 27 00:27:58 addons-305811 kubelet[2437]: E0927 00:27:58.684577 2437 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 33d3a52f94f9e6118a0621443bd1afb78a31eb5213ac2c2500cf78329a547aca" containerID="33d3a52f94f9e6118a0621443bd1afb78a31eb5213ac2c2500cf78329a547aca"
Sep 27 00:27:58 addons-305811 kubelet[2437]: I0927 00:27:58.684616 2437 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"33d3a52f94f9e6118a0621443bd1afb78a31eb5213ac2c2500cf78329a547aca"} err="failed to get container status \"33d3a52f94f9e6118a0621443bd1afb78a31eb5213ac2c2500cf78329a547aca\": rpc error: code = Unknown desc = Error response from daemon: No such container: 33d3a52f94f9e6118a0621443bd1afb78a31eb5213ac2c2500cf78329a547aca"
Sep 27 00:27:59 addons-305811 kubelet[2437]: I0927 00:27:59.043662 2437 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ff93bb-6e90-4ad1-8dc5-995f83e68be2" path="/var/lib/kubelet/pods/82ff93bb-6e90-4ad1-8dc5-995f83e68be2/volumes"
Sep 27 00:27:59 addons-305811 kubelet[2437]: I0927 00:27:59.043991 2437 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c5c71ea-d826-4aeb-a56d-116204ffb507" path="/var/lib/kubelet/pods/9c5c71ea-d826-4aeb-a56d-116204ffb507/volumes"
Sep 27 00:28:02 addons-305811 kubelet[2437]: I0927 00:28:02.658025 2437 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/dab9ce36-3057-4f92-8c1c-f9fa470c11a5-gcp-creds\") pod \"dab9ce36-3057-4f92-8c1c-f9fa470c11a5\" (UID: \"dab9ce36-3057-4f92-8c1c-f9fa470c11a5\") "
Sep 27 00:28:02 addons-305811 kubelet[2437]: I0927 00:28:02.658089 2437 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5z7s\" (UniqueName: \"kubernetes.io/projected/dab9ce36-3057-4f92-8c1c-f9fa470c11a5-kube-api-access-w5z7s\") pod \"dab9ce36-3057-4f92-8c1c-f9fa470c11a5\" (UID: \"dab9ce36-3057-4f92-8c1c-f9fa470c11a5\") "
Sep 27 00:28:02 addons-305811 kubelet[2437]: I0927 00:28:02.658147 2437 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dab9ce36-3057-4f92-8c1c-f9fa470c11a5-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "dab9ce36-3057-4f92-8c1c-f9fa470c11a5" (UID: "dab9ce36-3057-4f92-8c1c-f9fa470c11a5"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Sep 27 00:28:02 addons-305811 kubelet[2437]: I0927 00:28:02.659934 2437 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dab9ce36-3057-4f92-8c1c-f9fa470c11a5-kube-api-access-w5z7s" (OuterVolumeSpecName: "kube-api-access-w5z7s") pod "dab9ce36-3057-4f92-8c1c-f9fa470c11a5" (UID: "dab9ce36-3057-4f92-8c1c-f9fa470c11a5"). InnerVolumeSpecName "kube-api-access-w5z7s". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 27 00:28:02 addons-305811 kubelet[2437]: I0927 00:28:02.759261 2437 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/dab9ce36-3057-4f92-8c1c-f9fa470c11a5-gcp-creds\") on node \"addons-305811\" DevicePath \"\""
Sep 27 00:28:02 addons-305811 kubelet[2437]: I0927 00:28:02.759297 2437 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-w5z7s\" (UniqueName: \"kubernetes.io/projected/dab9ce36-3057-4f92-8c1c-f9fa470c11a5-kube-api-access-w5z7s\") on node \"addons-305811\" DevicePath \"\""
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.049068 2437 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dab9ce36-3057-4f92-8c1c-f9fa470c11a5" path="/var/lib/kubelet/pods/dab9ce36-3057-4f92-8c1c-f9fa470c11a5/volumes"
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.362911 2437 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdh4r\" (UniqueName: \"kubernetes.io/projected/4342257a-a438-4180-ab98-bcb513d0521a-kube-api-access-tdh4r\") pod \"4342257a-a438-4180-ab98-bcb513d0521a\" (UID: \"4342257a-a438-4180-ab98-bcb513d0521a\") "
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.362976 2437 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6bfc\" (UniqueName: \"kubernetes.io/projected/0fd59917-12f5-4b55-b4ed-fb31a0b82ca1-kube-api-access-v6bfc\") pod \"0fd59917-12f5-4b55-b4ed-fb31a0b82ca1\" (UID: \"0fd59917-12f5-4b55-b4ed-fb31a0b82ca1\") "
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.364928 2437 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fd59917-12f5-4b55-b4ed-fb31a0b82ca1-kube-api-access-v6bfc" (OuterVolumeSpecName: "kube-api-access-v6bfc") pod "0fd59917-12f5-4b55-b4ed-fb31a0b82ca1" (UID: "0fd59917-12f5-4b55-b4ed-fb31a0b82ca1"). InnerVolumeSpecName "kube-api-access-v6bfc". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.365012 2437 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4342257a-a438-4180-ab98-bcb513d0521a-kube-api-access-tdh4r" (OuterVolumeSpecName: "kube-api-access-tdh4r") pod "4342257a-a438-4180-ab98-bcb513d0521a" (UID: "4342257a-a438-4180-ab98-bcb513d0521a"). InnerVolumeSpecName "kube-api-access-tdh4r". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.463759 2437 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-v6bfc\" (UniqueName: \"kubernetes.io/projected/0fd59917-12f5-4b55-b4ed-fb31a0b82ca1-kube-api-access-v6bfc\") on node \"addons-305811\" DevicePath \"\""
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.463811 2437 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-tdh4r\" (UniqueName: \"kubernetes.io/projected/4342257a-a438-4180-ab98-bcb513d0521a-kube-api-access-tdh4r\") on node \"addons-305811\" DevicePath \"\""
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.721484 2437 scope.go:117] "RemoveContainer" containerID="0e1265bac22311a4cfc95dd200420ac264d53f97791ef81154582b8dac4eaf23"
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.739133 2437 scope.go:117] "RemoveContainer" containerID="0e1265bac22311a4cfc95dd200420ac264d53f97791ef81154582b8dac4eaf23"
Sep 27 00:28:03 addons-305811 kubelet[2437]: E0927 00:28:03.740115 2437 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 0e1265bac22311a4cfc95dd200420ac264d53f97791ef81154582b8dac4eaf23" containerID="0e1265bac22311a4cfc95dd200420ac264d53f97791ef81154582b8dac4eaf23"
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.740166 2437 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"0e1265bac22311a4cfc95dd200420ac264d53f97791ef81154582b8dac4eaf23"} err="failed to get container status \"0e1265bac22311a4cfc95dd200420ac264d53f97791ef81154582b8dac4eaf23\": rpc error: code = Unknown desc = Error response from daemon: No such container: 0e1265bac22311a4cfc95dd200420ac264d53f97791ef81154582b8dac4eaf23"
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.740248 2437 scope.go:117] "RemoveContainer" containerID="7ac1f7270abb14ba062978ccce74cfce81c4783b19231e4e36f8016eae0cf557"
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.761014 2437 scope.go:117] "RemoveContainer" containerID="7ac1f7270abb14ba062978ccce74cfce81c4783b19231e4e36f8016eae0cf557"
Sep 27 00:28:03 addons-305811 kubelet[2437]: E0927 00:28:03.762054 2437 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 7ac1f7270abb14ba062978ccce74cfce81c4783b19231e4e36f8016eae0cf557" containerID="7ac1f7270abb14ba062978ccce74cfce81c4783b19231e4e36f8016eae0cf557"
Sep 27 00:28:03 addons-305811 kubelet[2437]: I0927 00:28:03.762111 2437 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"7ac1f7270abb14ba062978ccce74cfce81c4783b19231e4e36f8016eae0cf557"} err="failed to get container status \"7ac1f7270abb14ba062978ccce74cfce81c4783b19231e4e36f8016eae0cf557\": rpc error: code = Unknown desc = Error response from daemon: No such container: 7ac1f7270abb14ba062978ccce74cfce81c4783b19231e4e36f8016eae0cf557"
==> storage-provisioner [5b2808a0c471] <==
I0927 00:15:35.230053 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0927 00:15:35.242455 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0927 00:15:35.242539 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0927 00:15:35.328280 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0927 00:15:35.328639 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-305811_59c1314c-8f98-4b2e-8d08-846e56086057!
I0927 00:15:35.331742 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"d7212221-b7e5-4ab5-9b0a-a78bfd6b7224", APIVersion:"v1", ResourceVersion:"603", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-305811_59c1314c-8f98-4b2e-8d08-846e56086057 became leader
I0927 00:15:35.429632 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-305811_59c1314c-8f98-4b2e-8d08-846e56086057!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-305811 -n addons-305811
helpers_test.go:261: (dbg) Run: kubectl --context addons-305811 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-305811 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-305811 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-305811/192.168.49.2
Start Time: Fri, 27 Sep 2024 00:18:49 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.27
IPs:
IP: 10.244.0.27
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-kfhm9 (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-kfhm9:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m15s default-scheduler Successfully assigned default/busybox to addons-305811
Normal Pulling 7m48s (x4 over 9m14s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m48s (x4 over 9m14s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m48s (x4 over 9m14s) kubelet Error: ErrImagePull
Warning Failed 7m22s (x6 over 9m14s) kubelet Error: ImagePullBackOff
Normal BackOff 4m2s (x20 over 9m14s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (72.54s)