=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:328: registry stabilized in 2.304235ms
addons_test.go:330: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-9bg4w" [da79db35-9dbe-40b6-bc10-153757b8bf2a] Running
addons_test.go:330: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.002563452s
addons_test.go:333: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-8lrkc" [0863352b-681f-45ef-a925-ee3ba3eb1198] Running
addons_test.go:333: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.002857322s
addons_test.go:338: (dbg) Run: kubectl --context addons-485025 delete po -l run=registry-test --now
addons_test.go:343: (dbg) Run: kubectl --context addons-485025 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:343: (dbg) Non-zero exit: kubectl --context addons-485025 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.075592196s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:345: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-485025 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:349: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:357: (dbg) Run: out/minikube-linux-amd64 -p addons-485025 ip
2024/09/30 10:33:53 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:386: (dbg) Run: out/minikube-linux-amd64 -p addons-485025 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-485025
helpers_test.go:235: (dbg) docker inspect addons-485025:
-- stdout --
[
{
"Id": "57c067488a26c5159be154c8352674e3f1d4a9cff700da00ad1c2b4e5cdb879d",
"Created": "2024-09-30T10:20:56.786096188Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 12517,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-30T10:20:56.915680776Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:fba5f082b59effd6acfcb1eed3d3f86a23bd3a65463877f8197a730d49f52a09",
"ResolvConfPath": "/var/lib/docker/containers/57c067488a26c5159be154c8352674e3f1d4a9cff700da00ad1c2b4e5cdb879d/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/57c067488a26c5159be154c8352674e3f1d4a9cff700da00ad1c2b4e5cdb879d/hostname",
"HostsPath": "/var/lib/docker/containers/57c067488a26c5159be154c8352674e3f1d4a9cff700da00ad1c2b4e5cdb879d/hosts",
"LogPath": "/var/lib/docker/containers/57c067488a26c5159be154c8352674e3f1d4a9cff700da00ad1c2b4e5cdb879d/57c067488a26c5159be154c8352674e3f1d4a9cff700da00ad1c2b4e5cdb879d-json.log",
"Name": "/addons-485025",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-485025:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-485025",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/1d326d47c67abcde5c405b5d2bc3203ddd9f5ed2ad55983bd8b9ac84aa3c1947-init/diff:/var/lib/docker/overlay2/71ddbbc874c8012f0ca6cba309f810cb206996525979cb1107bf2f7cf9f42c72/diff",
"MergedDir": "/var/lib/docker/overlay2/1d326d47c67abcde5c405b5d2bc3203ddd9f5ed2ad55983bd8b9ac84aa3c1947/merged",
"UpperDir": "/var/lib/docker/overlay2/1d326d47c67abcde5c405b5d2bc3203ddd9f5ed2ad55983bd8b9ac84aa3c1947/diff",
"WorkDir": "/var/lib/docker/overlay2/1d326d47c67abcde5c405b5d2bc3203ddd9f5ed2ad55983bd8b9ac84aa3c1947/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "volume",
"Name": "addons-485025",
"Source": "/var/lib/docker/volumes/addons-485025/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
},
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
}
],
"Config": {
"Hostname": "addons-485025",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-485025",
"name.minikube.sigs.k8s.io": "addons-485025",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "4c4a90f8c532aa8d22318a0057de685acfefe6d07f4992823fa5550e582622a4",
"SandboxKey": "/var/run/docker/netns/4c4a90f8c532",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-485025": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "a5e83fc3193b7ccbf4d12708117400c9967d758e8e666093fc0024b60a1253fc",
"EndpointID": "9aeabceb5b5d57a1b1623f8d8275166e161a072232abcdc94d3d91969c077b8b",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-485025",
"57c067488a26"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-485025 -n addons-485025
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-485025 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| start | --download-only -p | download-docker-079911 | jenkins | v1.34.0 | 30 Sep 24 10:20 UTC | |
| | download-docker-079911 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p download-docker-079911 | download-docker-079911 | jenkins | v1.34.0 | 30 Sep 24 10:20 UTC | 30 Sep 24 10:20 UTC |
| start | --download-only -p | binary-mirror-919884 | jenkins | v1.34.0 | 30 Sep 24 10:20 UTC | |
| | binary-mirror-919884 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:33823 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-919884 | binary-mirror-919884 | jenkins | v1.34.0 | 30 Sep 24 10:20 UTC | 30 Sep 24 10:20 UTC |
| addons | disable dashboard -p | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:20 UTC | |
| | addons-485025 | | | | | |
| addons | enable dashboard -p | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:20 UTC | |
| | addons-485025 | | | | | |
| start | -p addons-485025 --wait=true | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:20 UTC | 30 Sep 24 10:24 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| addons | addons-485025 addons disable | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:24 UTC | 30 Sep 24 10:24 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | enable headlamp | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:32 UTC | 30 Sep 24 10:32 UTC |
| | -p addons-485025 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-485025 addons | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:32 UTC | 30 Sep 24 10:32 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:32 UTC | 30 Sep 24 10:32 UTC |
| | addons-485025 | | | | | |
| addons | addons-485025 addons disable | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:32 UTC | 30 Sep 24 10:32 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ssh | addons-485025 ssh curl -s | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:32 UTC | 30 Sep 24 10:32 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-485025 ip | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:32 UTC | 30 Sep 24 10:32 UTC |
| addons | addons-485025 addons disable | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:32 UTC | 30 Sep 24 10:32 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-485025 addons disable | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:32 UTC | 30 Sep 24 10:33 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:33 UTC | 30 Sep 24 10:33 UTC |
| | -p addons-485025 | | | | | |
| addons | addons-485025 addons disable | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:33 UTC | 30 Sep 24 10:33 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| ssh | addons-485025 ssh cat | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:33 UTC | 30 Sep 24 10:33 UTC |
| | /opt/local-path-provisioner/pvc-c9d28883-8cdc-411a-b481-ed6040da0be1_default_test-pvc/file1 | | | | | |
| addons | addons-485025 addons disable | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:33 UTC | |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable cloud-spanner -p | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:33 UTC | 30 Sep 24 10:33 UTC |
| | addons-485025 | | | | | |
| addons | addons-485025 addons | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:33 UTC | 30 Sep 24 10:33 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-485025 addons | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:33 UTC | 30 Sep 24 10:33 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-485025 ip | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:33 UTC | 30 Sep 24 10:33 UTC |
| addons | addons-485025 addons disable | addons-485025 | jenkins | v1.34.0 | 30 Sep 24 10:33 UTC | 30 Sep 24 10:33 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/30 10:20:33
Running on machine: ubuntu-20-agent-14
Binary: Built with gc go1.23.0 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0930 10:20:33.219133 11756 out.go:345] Setting OutFile to fd 1 ...
I0930 10:20:33.219357 11756 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0930 10:20:33.219365 11756 out.go:358] Setting ErrFile to fd 2...
I0930 10:20:33.219369 11756 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0930 10:20:33.219541 11756 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19734-3685/.minikube/bin
I0930 10:20:33.220132 11756 out.go:352] Setting JSON to false
I0930 10:20:33.220967 11756 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":181,"bootTime":1727691452,"procs":180,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1069-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0930 10:20:33.221062 11756 start.go:139] virtualization: kvm guest
I0930 10:20:33.223169 11756 out.go:177] * [addons-485025] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0930 10:20:33.224434 11756 notify.go:220] Checking for updates...
I0930 10:20:33.224437 11756 out.go:177] - MINIKUBE_LOCATION=19734
I0930 10:20:33.225796 11756 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0930 10:20:33.227170 11756 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19734-3685/kubeconfig
I0930 10:20:33.228530 11756 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19734-3685/.minikube
I0930 10:20:33.229724 11756 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0930 10:20:33.231239 11756 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0930 10:20:33.232760 11756 driver.go:394] Setting default libvirt URI to qemu:///system
I0930 10:20:33.254767 11756 docker.go:123] docker version: linux-27.3.1:Docker Engine - Community
I0930 10:20:33.254851 11756 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0930 10:20:33.298501 11756 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:30 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-30 10:20:33.289418428 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.3.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:[WARNING: br
idge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled] ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.17.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.7] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0930 10:20:33.298646 11756 docker.go:318] overlay module found
I0930 10:20:33.301167 11756 out.go:177] * Using the docker driver based on user configuration
I0930 10:20:33.302222 11756 start.go:297] selected driver: docker
I0930 10:20:33.302234 11756 start.go:901] validating driver "docker" against <nil>
I0930 10:20:33.302244 11756 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0930 10:20:33.303029 11756 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0930 10:20:33.345538 11756 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-30 10:20:33.337483411 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.3.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:[WARNING: br
idge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled] ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.17.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.7] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0930 10:20:33.345676 11756 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0930 10:20:33.345911 11756 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0930 10:20:33.347559 11756 out.go:177] * Using Docker driver with root privileges
I0930 10:20:33.348677 11756 cni.go:84] Creating CNI manager for ""
I0930 10:20:33.348746 11756 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0930 10:20:33.348762 11756 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0930 10:20:33.348848 11756 start.go:340] cluster config:
{Name:addons-485025 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-485025 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0930 10:20:33.350140 11756 out.go:177] * Starting "addons-485025" primary control-plane node in "addons-485025" cluster
I0930 10:20:33.351199 11756 cache.go:121] Beginning downloading kic base image for docker with docker
I0930 10:20:33.352360 11756 out.go:177] * Pulling base image v0.0.45-1727108449-19696 ...
I0930 10:20:33.353610 11756 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0930 10:20:33.353640 11756 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19734-3685/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4
I0930 10:20:33.353647 11756 cache.go:56] Caching tarball of preloaded images
I0930 10:20:33.353697 11756 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 in local docker daemon
I0930 10:20:33.353716 11756 preload.go:172] Found /home/jenkins/minikube-integration/19734-3685/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0930 10:20:33.353723 11756 cache.go:59] Finished verifying existence of preloaded tar for v1.31.1 on docker
I0930 10:20:33.354029 11756 profile.go:143] Saving config to /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/config.json ...
I0930 10:20:33.354050 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/config.json: {Name:mk35b0d5ca357d92893ad556a0bad6107bb98cc5 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:20:33.368962 11756 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 to local cache
I0930 10:20:33.369057 11756 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 in local cache directory
I0930 10:20:33.369072 11756 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 in local cache directory, skipping pull
I0930 10:20:33.369076 11756 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 exists in cache, skipping pull
I0930 10:20:33.369085 11756 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 as a tarball
I0930 10:20:33.369092 11756 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 from local cache
I0930 10:20:45.093181 11756 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 from cached tarball
I0930 10:20:45.093220 11756 cache.go:194] Successfully downloaded all kic artifacts
I0930 10:20:45.093263 11756 start.go:360] acquireMachinesLock for addons-485025: {Name:mk599cf391d05d083ec36c01dacad090ed0c2f88 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0930 10:20:45.093351 11756 start.go:364] duration metric: took 69.396µs to acquireMachinesLock for "addons-485025"
I0930 10:20:45.093373 11756 start.go:93] Provisioning new machine with config: &{Name:addons-485025 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-485025 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0930 10:20:45.093454 11756 start.go:125] createHost starting for "" (driver="docker")
I0930 10:20:45.095092 11756 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0930 10:20:45.095325 11756 start.go:159] libmachine.API.Create for "addons-485025" (driver="docker")
I0930 10:20:45.095363 11756 client.go:168] LocalClient.Create starting
I0930 10:20:45.095475 11756 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19734-3685/.minikube/certs/ca.pem
I0930 10:20:45.343341 11756 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19734-3685/.minikube/certs/cert.pem
I0930 10:20:45.536576 11756 cli_runner.go:164] Run: docker network inspect addons-485025 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0930 10:20:45.551603 11756 cli_runner.go:211] docker network inspect addons-485025 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0930 10:20:45.551671 11756 network_create.go:284] running [docker network inspect addons-485025] to gather additional debugging logs...
I0930 10:20:45.551689 11756 cli_runner.go:164] Run: docker network inspect addons-485025
W0930 10:20:45.566238 11756 cli_runner.go:211] docker network inspect addons-485025 returned with exit code 1
I0930 10:20:45.566261 11756 network_create.go:287] error running [docker network inspect addons-485025]: docker network inspect addons-485025: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-485025 not found
I0930 10:20:45.566276 11756 network_create.go:289] output of [docker network inspect addons-485025]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-485025 not found
** /stderr **
I0930 10:20:45.566354 11756 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0930 10:20:45.581646 11756 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001500a60}
I0930 10:20:45.581689 11756 network_create.go:124] attempt to create docker network addons-485025 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0930 10:20:45.581732 11756 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-485025 addons-485025
I0930 10:20:45.640434 11756 network_create.go:108] docker network addons-485025 192.168.49.0/24 created
I0930 10:20:45.640530 11756 kic.go:121] calculated static IP "192.168.49.2" for the "addons-485025" container
I0930 10:20:45.640583 11756 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0930 10:20:45.655364 11756 cli_runner.go:164] Run: docker volume create addons-485025 --label name.minikube.sigs.k8s.io=addons-485025 --label created_by.minikube.sigs.k8s.io=true
I0930 10:20:45.671716 11756 oci.go:103] Successfully created a docker volume addons-485025
I0930 10:20:45.671776 11756 cli_runner.go:164] Run: docker run --rm --name addons-485025-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-485025 --entrypoint /usr/bin/test -v addons-485025:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 -d /var/lib
I0930 10:20:52.869918 11756 cli_runner.go:217] Completed: docker run --rm --name addons-485025-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-485025 --entrypoint /usr/bin/test -v addons-485025:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 -d /var/lib: (7.198105142s)
I0930 10:20:52.869942 11756 oci.go:107] Successfully prepared a docker volume addons-485025
I0930 10:20:52.869957 11756 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0930 10:20:52.869974 11756 kic.go:194] Starting extracting preloaded images to volume ...
I0930 10:20:52.870015 11756 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19734-3685/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-485025:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 -I lz4 -xf /preloaded.tar -C /extractDir
I0930 10:20:56.724866 11756 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19734-3685/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-485025:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 -I lz4 -xf /preloaded.tar -C /extractDir: (3.854797796s)
I0930 10:20:56.724902 11756 kic.go:203] duration metric: took 3.854924762s to extract preloaded images to volume ...
W0930 10:20:56.725054 11756 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0930 10:20:56.725191 11756 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0930 10:20:56.770876 11756 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-485025 --name addons-485025 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-485025 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-485025 --network addons-485025 --ip 192.168.49.2 --volume addons-485025:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21
I0930 10:20:57.086568 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Running}}
I0930 10:20:57.103912 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:20:57.121566 11756 cli_runner.go:164] Run: docker exec addons-485025 stat /var/lib/dpkg/alternatives/iptables
I0930 10:20:57.161487 11756 oci.go:144] the created container "addons-485025" has a running status.
I0930 10:20:57.161514 11756 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa...
I0930 10:20:57.413896 11756 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0930 10:20:57.433465 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:20:57.453562 11756 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0930 10:20:57.453591 11756 kic_runner.go:114] Args: [docker exec --privileged addons-485025 chown docker:docker /home/docker/.ssh/authorized_keys]
I0930 10:20:57.492949 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:20:57.513671 11756 machine.go:93] provisionDockerMachine start ...
I0930 10:20:57.513769 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:57.533177 11756 main.go:141] libmachine: Using SSH client type: native
I0930 10:20:57.533391 11756 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0930 10:20:57.533407 11756 main.go:141] libmachine: About to run SSH command:
hostname
I0930 10:20:57.659673 11756 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-485025
I0930 10:20:57.659698 11756 ubuntu.go:169] provisioning hostname "addons-485025"
I0930 10:20:57.659756 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:57.677329 11756 main.go:141] libmachine: Using SSH client type: native
I0930 10:20:57.677522 11756 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0930 10:20:57.677544 11756 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-485025 && echo "addons-485025" | sudo tee /etc/hostname
I0930 10:20:57.798265 11756 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-485025
I0930 10:20:57.798336 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:57.814131 11756 main.go:141] libmachine: Using SSH client type: native
I0930 10:20:57.814300 11756 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0930 10:20:57.814317 11756 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-485025' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-485025/g' /etc/hosts;
else
echo '127.0.1.1 addons-485025' | sudo tee -a /etc/hosts;
fi
fi
I0930 10:20:57.928277 11756 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0930 10:20:57.928307 11756 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19734-3685/.minikube CaCertPath:/home/jenkins/minikube-integration/19734-3685/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19734-3685/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19734-3685/.minikube}
I0930 10:20:57.928367 11756 ubuntu.go:177] setting up certificates
I0930 10:20:57.928379 11756 provision.go:84] configureAuth start
I0930 10:20:57.928421 11756 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-485025
I0930 10:20:57.944024 11756 provision.go:143] copyHostCerts
I0930 10:20:57.944086 11756 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19734-3685/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19734-3685/.minikube/ca.pem (1078 bytes)
I0930 10:20:57.944206 11756 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19734-3685/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19734-3685/.minikube/cert.pem (1123 bytes)
I0930 10:20:57.944279 11756 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19734-3685/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19734-3685/.minikube/key.pem (1675 bytes)
I0930 10:20:57.944356 11756 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19734-3685/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19734-3685/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19734-3685/.minikube/certs/ca-key.pem org=jenkins.addons-485025 san=[127.0.0.1 192.168.49.2 addons-485025 localhost minikube]
I0930 10:20:58.113651 11756 provision.go:177] copyRemoteCerts
I0930 10:20:58.113705 11756 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0930 10:20:58.113738 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:58.129515 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:20:58.212861 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0930 10:20:58.234883 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0930 10:20:58.255499 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0930 10:20:58.275355 11756 provision.go:87] duration metric: took 346.962396ms to configureAuth
I0930 10:20:58.275390 11756 ubuntu.go:193] setting minikube options for container-runtime
I0930 10:20:58.275546 11756 config.go:182] Loaded profile config "addons-485025": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0930 10:20:58.275597 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:58.292012 11756 main.go:141] libmachine: Using SSH client type: native
I0930 10:20:58.292177 11756 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0930 10:20:58.292189 11756 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0930 10:20:58.404299 11756 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0930 10:20:58.404345 11756 ubuntu.go:71] root file system type: overlay
I0930 10:20:58.404508 11756 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0930 10:20:58.404585 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:58.420748 11756 main.go:141] libmachine: Using SSH client type: native
I0930 10:20:58.420915 11756 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0930 10:20:58.420972 11756 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0930 10:20:58.542243 11756 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0930 10:20:58.542323 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:58.558317 11756 main.go:141] libmachine: Using SSH client type: native
I0930 10:20:58.558477 11756 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0930 10:20:58.558495 11756 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0930 10:20:59.220674 11756 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-09-20 11:39:29.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-30 10:20:58.537302378 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0930 10:20:59.220706 11756 machine.go:96] duration metric: took 1.707008832s to provisionDockerMachine
I0930 10:20:59.220719 11756 client.go:171] duration metric: took 14.125345483s to LocalClient.Create
I0930 10:20:59.220741 11756 start.go:167] duration metric: took 14.125415591s to libmachine.API.Create "addons-485025"
I0930 10:20:59.220751 11756 start.go:293] postStartSetup for "addons-485025" (driver="docker")
I0930 10:20:59.220761 11756 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0930 10:20:59.220819 11756 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0930 10:20:59.220859 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:59.237212 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:20:59.320350 11756 ssh_runner.go:195] Run: cat /etc/os-release
I0930 10:20:59.323087 11756 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0930 10:20:59.323114 11756 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0930 10:20:59.323125 11756 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0930 10:20:59.323130 11756 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0930 10:20:59.323141 11756 filesync.go:126] Scanning /home/jenkins/minikube-integration/19734-3685/.minikube/addons for local assets ...
I0930 10:20:59.323195 11756 filesync.go:126] Scanning /home/jenkins/minikube-integration/19734-3685/.minikube/files for local assets ...
I0930 10:20:59.323217 11756 start.go:296] duration metric: took 102.46109ms for postStartSetup
I0930 10:20:59.323466 11756 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-485025
I0930 10:20:59.338761 11756 profile.go:143] Saving config to /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/config.json ...
I0930 10:20:59.338995 11756 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0930 10:20:59.339031 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:59.354357 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:20:59.432660 11756 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0930 10:20:59.436441 11756 start.go:128] duration metric: took 14.342975669s to createHost
I0930 10:20:59.436464 11756 start.go:83] releasing machines lock for "addons-485025", held for 14.343101517s
I0930 10:20:59.436517 11756 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-485025
I0930 10:20:59.452089 11756 ssh_runner.go:195] Run: cat /version.json
I0930 10:20:59.452131 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:59.452176 11756 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0930 10:20:59.452243 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:20:59.468987 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:20:59.469635 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:20:59.614213 11756 ssh_runner.go:195] Run: systemctl --version
I0930 10:20:59.617896 11756 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0930 10:20:59.621440 11756 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0930 10:20:59.641760 11756 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0930 10:20:59.641816 11756 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0930 10:20:59.664706 11756 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0930 10:20:59.664727 11756 start.go:495] detecting cgroup driver to use...
I0930 10:20:59.664761 11756 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0930 10:20:59.664855 11756 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0930 10:20:59.677870 11756 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0930 10:20:59.685609 11756 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0930 10:20:59.693489 11756 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0930 10:20:59.693550 11756 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0930 10:20:59.701321 11756 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0930 10:20:59.709250 11756 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0930 10:20:59.717110 11756 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0930 10:20:59.725114 11756 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0930 10:20:59.732630 11756 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0930 10:20:59.740614 11756 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0930 10:20:59.748687 11756 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0930 10:20:59.756493 11756 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0930 10:20:59.763258 11756 crio.go:166] couldn't verify netfilter by "sudo sysctl net.bridge.bridge-nf-call-iptables" which might be okay. error: sudo sysctl net.bridge.bridge-nf-call-iptables: Process exited with status 255
stdout:
stderr:
sysctl: cannot stat /proc/sys/net/bridge/bridge-nf-call-iptables: No such file or directory
I0930 10:20:59.763300 11756 ssh_runner.go:195] Run: sudo modprobe br_netfilter
I0930 10:20:59.774506 11756 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0930 10:20:59.781115 11756 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0930 10:20:59.849023 11756 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0930 10:20:59.921415 11756 start.go:495] detecting cgroup driver to use...
I0930 10:20:59.921479 11756 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0930 10:20:59.921523 11756 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0930 10:20:59.931848 11756 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0930 10:20:59.931915 11756 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0930 10:20:59.943115 11756 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0930 10:20:59.958258 11756 ssh_runner.go:195] Run: which cri-dockerd
I0930 10:20:59.961374 11756 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0930 10:20:59.970585 11756 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0930 10:20:59.988236 11756 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0930 10:21:00.087140 11756 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0930 10:21:00.177813 11756 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0930 10:21:00.177931 11756 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0930 10:21:00.193665 11756 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0930 10:21:00.265382 11756 ssh_runner.go:195] Run: sudo systemctl restart docker
I0930 10:21:00.498726 11756 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0930 10:21:00.508842 11756 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0930 10:21:00.518287 11756 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0930 10:21:00.593272 11756 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0930 10:21:00.664526 11756 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0930 10:21:00.738087 11756 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0930 10:21:00.749792 11756 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0930 10:21:00.759287 11756 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0930 10:21:00.830699 11756 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0930 10:21:00.888454 11756 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0930 10:21:00.888529 11756 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0930 10:21:00.891663 11756 start.go:563] Will wait 60s for crictl version
I0930 10:21:00.891714 11756 ssh_runner.go:195] Run: which crictl
I0930 10:21:00.894654 11756 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0930 10:21:00.924382 11756 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.3.1
RuntimeApiVersion: v1
I0930 10:21:00.924448 11756 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0930 10:21:00.946157 11756 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0930 10:21:00.970224 11756 out.go:235] * Preparing Kubernetes v1.31.1 on Docker 27.3.1 ...
I0930 10:21:00.970298 11756 cli_runner.go:164] Run: docker network inspect addons-485025 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0930 10:21:00.985001 11756 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0930 10:21:00.988197 11756 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0930 10:21:00.997681 11756 kubeadm.go:883] updating cluster {Name:addons-485025 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-485025 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0930 10:21:00.997780 11756 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0930 10:21:00.997823 11756 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0930 10:21:01.016460 11756 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0930 10:21:01.016492 11756 docker.go:615] Images already preloaded, skipping extraction
I0930 10:21:01.016542 11756 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0930 10:21:01.033501 11756 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0930 10:21:01.033528 11756 cache_images.go:84] Images are preloaded, skipping loading
I0930 10:21:01.033537 11756 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.1 docker true true} ...
I0930 10:21:01.033622 11756 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-485025 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.1 ClusterName:addons-485025 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0930 10:21:01.033668 11756 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0930 10:21:01.074112 11756 cni.go:84] Creating CNI manager for ""
I0930 10:21:01.074135 11756 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0930 10:21:01.074144 11756 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0930 10:21:01.074161 11756 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-485025 NodeName:addons-485025 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0930 10:21:01.074275 11756 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-485025"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0930 10:21:01.074323 11756 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.1
I0930 10:21:01.082038 11756 binaries.go:44] Found k8s binaries, skipping transfer
I0930 10:21:01.082091 11756 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0930 10:21:01.089518 11756 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0930 10:21:01.104124 11756 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0930 10:21:01.118583 11756 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0930 10:21:01.133036 11756 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0930 10:21:01.135706 11756 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0930 10:21:01.144832 11756 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0930 10:21:01.224247 11756 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0930 10:21:01.236076 11756 certs.go:68] Setting up /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025 for IP: 192.168.49.2
I0930 10:21:01.236094 11756 certs.go:194] generating shared ca certs ...
I0930 10:21:01.236107 11756 certs.go:226] acquiring lock for ca certs: {Name:mk681cd5e73e48fcc7a587a82627f61623810efe Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.236206 11756 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19734-3685/.minikube/ca.key
I0930 10:21:01.334683 11756 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19734-3685/.minikube/ca.crt ...
I0930 10:21:01.334710 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/ca.crt: {Name:mkdee3312b387c39866499d029665cc3f900e216 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.334866 11756 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19734-3685/.minikube/ca.key ...
I0930 10:21:01.334876 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/ca.key: {Name:mk407f9207954f7156758dee60d9450547f464a8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.334946 11756 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19734-3685/.minikube/proxy-client-ca.key
I0930 10:21:01.424081 11756 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19734-3685/.minikube/proxy-client-ca.crt ...
I0930 10:21:01.424106 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/proxy-client-ca.crt: {Name:mk7a32903d7331d5a1ac8dd98f3fadc8e85d608d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.424246 11756 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19734-3685/.minikube/proxy-client-ca.key ...
I0930 10:21:01.424256 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/proxy-client-ca.key: {Name:mk908174bbb25f270b91f85987a291c59ffcb1a7 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.424314 11756 certs.go:256] generating profile certs ...
I0930 10:21:01.424385 11756 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/client.key
I0930 10:21:01.424402 11756 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/client.crt with IP's: []
I0930 10:21:01.600394 11756 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/client.crt ...
I0930 10:21:01.600420 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/client.crt: {Name:mk83ac12329b857a1fef1eebd94b6263c29b0ad8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.600570 11756 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/client.key ...
I0930 10:21:01.600579 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/client.key: {Name:mk63374181f78b45cccbae715d9b62677444f222 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.600644 11756 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.key.301cc93f
I0930 10:21:01.600662 11756 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.crt.301cc93f with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0930 10:21:01.732610 11756 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.crt.301cc93f ...
I0930 10:21:01.732639 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.crt.301cc93f: {Name:mk7dabdde80e57d85cae52592cdbcbb2db8c842e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.732785 11756 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.key.301cc93f ...
I0930 10:21:01.732795 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.key.301cc93f: {Name:mk76c6232d9c7f5efdefb49845b7c304ace06c0f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.732868 11756 certs.go:381] copying /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.crt.301cc93f -> /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.crt
I0930 10:21:01.732937 11756 certs.go:385] copying /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.key.301cc93f -> /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.key
I0930 10:21:01.732981 11756 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/proxy-client.key
I0930 10:21:01.732999 11756 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/proxy-client.crt with IP's: []
I0930 10:21:01.879796 11756 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/proxy-client.crt ...
I0930 10:21:01.879836 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/proxy-client.crt: {Name:mk418eeddcbb7b19c4f94e41b568c7e6aae3a678 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.880040 11756 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/proxy-client.key ...
I0930 10:21:01.880056 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/proxy-client.key: {Name:mk93e9608c5e44aa9cfe3d6336fe8342484835a5 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:01.880257 11756 certs.go:484] found cert: /home/jenkins/minikube-integration/19734-3685/.minikube/certs/ca-key.pem (1679 bytes)
I0930 10:21:01.880303 11756 certs.go:484] found cert: /home/jenkins/minikube-integration/19734-3685/.minikube/certs/ca.pem (1078 bytes)
I0930 10:21:01.880362 11756 certs.go:484] found cert: /home/jenkins/minikube-integration/19734-3685/.minikube/certs/cert.pem (1123 bytes)
I0930 10:21:01.880397 11756 certs.go:484] found cert: /home/jenkins/minikube-integration/19734-3685/.minikube/certs/key.pem (1675 bytes)
I0930 10:21:01.881047 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0930 10:21:01.902522 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0930 10:21:01.922581 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0930 10:21:01.942764 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0930 10:21:01.962610 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0930 10:21:01.982652 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0930 10:21:02.003721 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0930 10:21:02.025021 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/profiles/addons-485025/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0930 10:21:02.046049 11756 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19734-3685/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0930 10:21:02.065988 11756 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0930 10:21:02.080873 11756 ssh_runner.go:195] Run: openssl version
I0930 10:21:02.085622 11756 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0930 10:21:02.093399 11756 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0930 10:21:02.096334 11756 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 30 10:21 /usr/share/ca-certificates/minikubeCA.pem
I0930 10:21:02.096389 11756 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0930 10:21:02.102353 11756 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0930 10:21:02.110105 11756 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0930 10:21:02.112640 11756 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0930 10:21:02.112682 11756 kubeadm.go:392] StartCluster: {Name:addons-485025 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1727108449-19696@sha256:c662152d8855bc4c62a3b5786a68adf99e04794e7f8f374a3859703004ef1d21 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-485025 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0930 10:21:02.112778 11756 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0930 10:21:02.128932 11756 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0930 10:21:02.136471 11756 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0930 10:21:02.143843 11756 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0930 10:21:02.143888 11756 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0930 10:21:02.151112 11756 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0930 10:21:02.151131 11756 kubeadm.go:157] found existing configuration files:
I0930 10:21:02.151164 11756 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0930 10:21:02.158226 11756 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0930 10:21:02.158280 11756 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0930 10:21:02.165155 11756 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0930 10:21:02.172227 11756 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0930 10:21:02.172273 11756 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0930 10:21:02.178889 11756 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0930 10:21:02.185637 11756 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0930 10:21:02.185686 11756 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0930 10:21:02.192143 11756 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0930 10:21:02.198874 11756 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0930 10:21:02.198907 11756 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0930 10:21:02.205471 11756 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0930 10:21:02.238813 11756 kubeadm.go:310] [init] Using Kubernetes version: v1.31.1
I0930 10:21:02.238872 11756 kubeadm.go:310] [preflight] Running pre-flight checks
I0930 10:21:02.257143 11756 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0930 10:21:02.257214 11756 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1069-gcp[0m
I0930 10:21:02.257251 11756 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0930 10:21:02.257308 11756 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0930 10:21:02.257358 11756 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0930 10:21:02.257411 11756 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0930 10:21:02.257468 11756 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0930 10:21:02.257518 11756 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0930 10:21:02.257612 11756 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0930 10:21:02.257656 11756 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0930 10:21:02.257726 11756 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0930 10:21:02.257804 11756 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0930 10:21:02.304155 11756 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0930 10:21:02.304367 11756 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0930 10:21:02.304515 11756 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0930 10:21:02.314270 11756 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0930 10:21:02.317649 11756 out.go:235] - Generating certificates and keys ...
I0930 10:21:02.317750 11756 kubeadm.go:310] [certs] Using existing ca certificate authority
I0930 10:21:02.317850 11756 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0930 10:21:02.486398 11756 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0930 10:21:02.589470 11756 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0930 10:21:02.670718 11756 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0930 10:21:02.785854 11756 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0930 10:21:02.829563 11756 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0930 10:21:02.829697 11756 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-485025 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0930 10:21:02.950949 11756 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0930 10:21:02.951122 11756 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-485025 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0930 10:21:03.128508 11756 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0930 10:21:03.307756 11756 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0930 10:21:03.515121 11756 kubeadm.go:310] [certs] Generating "sa" key and public key
I0930 10:21:03.515207 11756 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0930 10:21:03.826089 11756 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0930 10:21:04.065847 11756 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0930 10:21:04.320280 11756 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0930 10:21:04.775424 11756 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0930 10:21:04.855255 11756 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0930 10:21:04.855849 11756 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0930 10:21:04.858139 11756 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0930 10:21:04.860027 11756 out.go:235] - Booting up control plane ...
I0930 10:21:04.860124 11756 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0930 10:21:04.860215 11756 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0930 10:21:04.860313 11756 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0930 10:21:04.868681 11756 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0930 10:21:04.873347 11756 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0930 10:21:04.873415 11756 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0930 10:21:04.952159 11756 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0930 10:21:04.952355 11756 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0930 10:21:05.453403 11756 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 501.466407ms
I0930 10:21:05.453493 11756 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0930 10:21:09.954613 11756 kubeadm.go:310] [api-check] The API server is healthy after 4.501160526s
I0930 10:21:09.964808 11756 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0930 10:21:09.973906 11756 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0930 10:21:09.991429 11756 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0930 10:21:09.991672 11756 kubeadm.go:310] [mark-control-plane] Marking the node addons-485025 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0930 10:21:09.998108 11756 kubeadm.go:310] [bootstrap-token] Using token: f4zrib.168z9svcsibzej7f
I0930 10:21:09.999453 11756 out.go:235] - Configuring RBAC rules ...
I0930 10:21:09.999625 11756 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0930 10:21:10.002377 11756 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0930 10:21:10.009745 11756 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0930 10:21:10.011812 11756 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0930 10:21:10.014068 11756 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0930 10:21:10.016185 11756 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0930 10:21:10.359920 11756 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0930 10:21:10.775527 11756 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0930 10:21:11.361158 11756 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0930 10:21:11.361970 11756 kubeadm.go:310]
I0930 10:21:11.362066 11756 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0930 10:21:11.362078 11756 kubeadm.go:310]
I0930 10:21:11.362179 11756 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0930 10:21:11.362189 11756 kubeadm.go:310]
I0930 10:21:11.362238 11756 kubeadm.go:310] mkdir -p $HOME/.kube
I0930 10:21:11.362325 11756 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0930 10:21:11.362408 11756 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0930 10:21:11.362417 11756 kubeadm.go:310]
I0930 10:21:11.362512 11756 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0930 10:21:11.362538 11756 kubeadm.go:310]
I0930 10:21:11.362622 11756 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0930 10:21:11.362632 11756 kubeadm.go:310]
I0930 10:21:11.362706 11756 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0930 10:21:11.362829 11756 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0930 10:21:11.362926 11756 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0930 10:21:11.362935 11756 kubeadm.go:310]
I0930 10:21:11.363042 11756 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0930 10:21:11.363148 11756 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0930 10:21:11.363157 11756 kubeadm.go:310]
I0930 10:21:11.363265 11756 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token f4zrib.168z9svcsibzej7f \
I0930 10:21:11.363392 11756 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:2019e7f4cdd44306d6ad5bfe800e8b084e0bcb230a13ba581c51d5d41d39980c \
I0930 10:21:11.363427 11756 kubeadm.go:310] --control-plane
I0930 10:21:11.363435 11756 kubeadm.go:310]
I0930 10:21:11.363536 11756 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0930 10:21:11.363545 11756 kubeadm.go:310]
I0930 10:21:11.363628 11756 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token f4zrib.168z9svcsibzej7f \
I0930 10:21:11.363726 11756 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:2019e7f4cdd44306d6ad5bfe800e8b084e0bcb230a13ba581c51d5d41d39980c
I0930 10:21:11.365434 11756 kubeadm.go:310] W0930 10:21:02.236453 1926 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0930 10:21:11.365694 11756 kubeadm.go:310] W0930 10:21:02.237051 1926 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0930 10:21:11.365892 11756 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1069-gcp\n", err: exit status 1
I0930 10:21:11.366060 11756 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0930 10:21:11.366091 11756 cni.go:84] Creating CNI manager for ""
I0930 10:21:11.366112 11756 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0930 10:21:11.367633 11756 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0930 10:21:11.368808 11756 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0930 10:21:11.376705 11756 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0930 10:21:11.392399 11756 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0930 10:21:11.392505 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:11.392528 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-485025 minikube.k8s.io/updated_at=2024_09_30T10_21_11_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=b68b4b088317c82ffa16da1c47933e77f0f5d128 minikube.k8s.io/name=addons-485025 minikube.k8s.io/primary=true
I0930 10:21:11.399034 11756 ops.go:34] apiserver oom_adj: -16
I0930 10:21:11.463873 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:11.964838 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:12.464447 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:12.964756 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:13.464510 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:13.964073 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:14.464665 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:14.964446 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:15.463961 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:15.964053 11756 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0930 10:21:16.022888 11756 kubeadm.go:1113] duration metric: took 4.630450401s to wait for elevateKubeSystemPrivileges
I0930 10:21:16.022926 11756 kubeadm.go:394] duration metric: took 13.910246458s to StartCluster
I0930 10:21:16.022942 11756 settings.go:142] acquiring lock: {Name:mk72471fb7cd04ec1061860566590829aa9a0fa4 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:16.023038 11756 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19734-3685/kubeconfig
I0930 10:21:16.023352 11756 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19734-3685/kubeconfig: {Name:mkb502c0ffb71c0d28ec7c189a15446320a80b42 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0930 10:21:16.023509 11756 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0930 10:21:16.023554 11756 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0930 10:21:16.023621 11756 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0930 10:21:16.023737 11756 addons.go:69] Setting yakd=true in profile "addons-485025"
I0930 10:21:16.023759 11756 addons.go:234] Setting addon yakd=true in "addons-485025"
I0930 10:21:16.023754 11756 addons.go:69] Setting ingress=true in profile "addons-485025"
I0930 10:21:16.023792 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.023796 11756 addons.go:234] Setting addon ingress=true in "addons-485025"
I0930 10:21:16.023801 11756 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-485025"
I0930 10:21:16.023800 11756 addons.go:69] Setting inspektor-gadget=true in profile "addons-485025"
I0930 10:21:16.023817 11756 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-485025"
I0930 10:21:16.023827 11756 addons.go:234] Setting addon inspektor-gadget=true in "addons-485025"
I0930 10:21:16.023830 11756 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-485025"
I0930 10:21:16.023833 11756 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-485025"
I0930 10:21:16.023843 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.023856 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.023867 11756 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-485025"
I0930 10:21:16.023881 11756 addons.go:69] Setting storage-provisioner=true in profile "addons-485025"
I0930 10:21:16.023899 11756 addons.go:234] Setting addon storage-provisioner=true in "addons-485025"
I0930 10:21:16.023908 11756 addons.go:69] Setting cloud-spanner=true in profile "addons-485025"
I0930 10:21:16.023920 11756 addons.go:234] Setting addon cloud-spanner=true in "addons-485025"
I0930 10:21:16.023924 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.023938 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.024151 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.024279 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.024306 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.024341 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.024373 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.023862 11756 addons.go:69] Setting default-storageclass=true in profile "addons-485025"
I0930 10:21:16.024383 11756 addons.go:69] Setting gcp-auth=true in profile "addons-485025"
I0930 10:21:16.024433 11756 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-485025"
I0930 10:21:16.024439 11756 mustload.go:65] Loading cluster: addons-485025
I0930 10:21:16.023791 11756 addons.go:69] Setting metrics-server=true in profile "addons-485025"
I0930 10:21:16.024510 11756 addons.go:234] Setting addon metrics-server=true in "addons-485025"
I0930 10:21:16.024532 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.024599 11756 config.go:182] Loaded profile config "addons-485025": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0930 10:21:16.024700 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.023797 11756 config.go:182] Loaded profile config "addons-485025": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0930 10:21:16.024403 11756 addons.go:69] Setting ingress-dns=true in profile "addons-485025"
I0930 10:21:16.024770 11756 addons.go:234] Setting addon ingress-dns=true in "addons-485025"
I0930 10:21:16.024800 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.024821 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.024984 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.025271 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.023768 11756 addons.go:69] Setting registry=true in profile "addons-485025"
I0930 10:21:16.025545 11756 addons.go:234] Setting addon registry=true in "addons-485025"
I0930 10:21:16.025585 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.026065 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.023819 11756 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-485025"
I0930 10:21:16.027061 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.024396 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.024414 11756 addons.go:69] Setting volcano=true in profile "addons-485025"
I0930 10:21:16.030592 11756 addons.go:234] Setting addon volcano=true in "addons-485025"
I0930 10:21:16.030689 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.030802 11756 out.go:177] * Verifying Kubernetes components...
I0930 10:21:16.024416 11756 addons.go:69] Setting volumesnapshots=true in profile "addons-485025"
I0930 10:21:16.031623 11756 addons.go:234] Setting addon volumesnapshots=true in "addons-485025"
I0930 10:21:16.031674 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.032125 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.023900 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.032671 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.032849 11756 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0930 10:21:16.033192 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.057124 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.074152 11756 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0930 10:21:16.074954 11756 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0930 10:21:16.075759 11756 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0930 10:21:16.075782 11756 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0930 10:21:16.075862 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.076640 11756 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0930 10:21:16.076658 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0930 10:21:16.076712 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.081840 11756 addons.go:234] Setting addon default-storageclass=true in "addons-485025"
I0930 10:21:16.081884 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.082319 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.082744 11756 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0930 10:21:16.084865 11756 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0930 10:21:16.084891 11756 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0930 10:21:16.084944 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.085355 11756 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-485025"
I0930 10:21:16.085408 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.085936 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:16.090976 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:16.095375 11756 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0930 10:21:16.096367 11756 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0930 10:21:16.096447 11756 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0930 10:21:16.096464 11756 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0930 10:21:16.096523 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.097487 11756 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0930 10:21:16.097508 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0930 10:21:16.097560 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.112714 11756 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.7
I0930 10:21:16.114984 11756 out.go:177] - Using image docker.io/registry:2.8.3
I0930 10:21:16.116579 11756 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0930 10:21:16.116604 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0930 10:21:16.116739 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.123296 11756 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0930 10:21:16.128447 11756 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0930 10:21:16.128474 11756 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0930 10:21:16.128532 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.130281 11756 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0930 10:21:16.131789 11756 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0930 10:21:16.133184 11756 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0930 10:21:16.133207 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0930 10:21:16.133263 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.138718 11756 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0930 10:21:16.139765 11756 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0930 10:21:16.139783 11756 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0930 10:21:16.139808 11756 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0930 10:21:16.139857 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.141977 11756 out.go:177] - Using image docker.io/busybox:stable
I0930 10:21:16.143281 11756 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0930 10:21:16.143300 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0930 10:21:16.143351 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.146004 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.146530 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.147825 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.149089 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.151912 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.164479 11756 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.10.0
I0930 10:21:16.164550 11756 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0930 10:21:16.166611 11756 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0930 10:21:16.166626 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0930 10:21:16.166681 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.166841 11756 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.10.0
I0930 10:21:16.168301 11756 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.10.0
I0930 10:21:16.170798 11756 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0930 10:21:16.170821 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (471825 bytes)
I0930 10:21:16.170870 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.170963 11756 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0930 10:21:16.172529 11756 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0930 10:21:16.173576 11756 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0930 10:21:16.174843 11756 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0930 10:21:16.176145 11756 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0930 10:21:16.177434 11756 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0930 10:21:16.178455 11756 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0930 10:21:16.179540 11756 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0930 10:21:16.180683 11756 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0930 10:21:16.180705 11756 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0930 10:21:16.180759 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.181121 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.181281 11756 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.24
I0930 10:21:16.186298 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.186541 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.188289 11756 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0930 10:21:16.188303 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0930 10:21:16.188407 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:16.199205 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.204726 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.204731 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.214366 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.218319 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:16.220951 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
W0930 10:21:16.252637 11756 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0930 10:21:16.252669 11756 retry.go:31] will retry after 232.123848ms: ssh: handshake failed: EOF
I0930 10:21:16.451637 11756 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0930 10:21:16.455002 11756 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0930 10:21:16.550773 11756 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0930 10:21:16.550797 11756 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0930 10:21:16.573602 11756 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0930 10:21:16.573692 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0930 10:21:16.574147 11756 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0930 10:21:16.574166 11756 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0930 10:21:16.649864 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0930 10:21:16.658600 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0930 10:21:16.762605 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0930 10:21:16.764584 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0930 10:21:16.765559 11756 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0930 10:21:16.765655 11756 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0930 10:21:16.771310 11756 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0930 10:21:16.771365 11756 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0930 10:21:16.850159 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0930 10:21:16.851243 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0930 10:21:16.861569 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0930 10:21:16.865447 11756 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0930 10:21:16.865497 11756 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0930 10:21:16.954593 11756 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0930 10:21:16.954635 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0930 10:21:16.954905 11756 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0930 10:21:16.954926 11756 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0930 10:21:16.958759 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0930 10:21:17.061614 11756 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0930 10:21:17.061703 11756 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0930 10:21:17.072970 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0930 10:21:17.170637 11756 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0930 10:21:17.170717 11756 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0930 10:21:17.256054 11756 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0930 10:21:17.256086 11756 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0930 10:21:17.265029 11756 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0930 10:21:17.265055 11756 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0930 10:21:17.351890 11756 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0930 10:21:17.351921 11756 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0930 10:21:17.450182 11756 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0930 10:21:17.450208 11756 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0930 10:21:17.470612 11756 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0930 10:21:17.470689 11756 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0930 10:21:17.754449 11756 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0930 10:21:17.754489 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0930 10:21:17.855454 11756 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0930 10:21:17.855483 11756 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0930 10:21:17.966172 11756 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0930 10:21:17.966202 11756 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0930 10:21:18.049700 11756 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0930 10:21:18.049741 11756 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0930 10:21:18.056418 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0930 10:21:18.157036 11756 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.705336074s)
I0930 10:21:18.157133 11756 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0930 10:21:18.157334 11756 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.702306884s)
I0930 10:21:18.159295 11756 node_ready.go:35] waiting up to 6m0s for node "addons-485025" to be "Ready" ...
I0930 10:21:18.163729 11756 node_ready.go:49] node "addons-485025" has status "Ready":"True"
I0930 10:21:18.163751 11756 node_ready.go:38] duration metric: took 4.387973ms for node "addons-485025" to be "Ready" ...
I0930 10:21:18.163760 11756 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0930 10:21:18.173013 11756 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-qxcw9" in "kube-system" namespace to be "Ready" ...
I0930 10:21:18.257187 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0930 10:21:18.369427 11756 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0930 10:21:18.369518 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0930 10:21:18.550480 11756 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0930 10:21:18.550583 11756 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0930 10:21:18.663508 11756 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-485025" context rescaled to 1 replicas
I0930 10:21:18.765168 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0930 10:21:18.949982 11756 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0930 10:21:18.950009 11756 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0930 10:21:19.452593 11756 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0930 10:21:19.452631 11756 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0930 10:21:19.567372 11756 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0930 10:21:19.567463 11756 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0930 10:21:19.663756 11756 addons.go:431] installing /etc/kubernetes/addons/ig-configmap.yaml
I0930 10:21:19.663833 11756 ssh_runner.go:362] scp inspektor-gadget/ig-configmap.yaml --> /etc/kubernetes/addons/ig-configmap.yaml (754 bytes)
I0930 10:21:19.852285 11756 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0930 10:21:19.852402 11756 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0930 10:21:20.258294 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-qxcw9" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:20.259118 11756 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0930 10:21:20.259141 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (8196 bytes)
I0930 10:21:20.350487 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (3.700581366s)
I0930 10:21:20.350614 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (3.691990695s)
I0930 10:21:20.356247 11756 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0930 10:21:20.356272 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0930 10:21:20.562181 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-configmap.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0930 10:21:20.953868 11756 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0930 10:21:20.953957 11756 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0930 10:21:21.271931 11756 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0930 10:21:21.271957 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0930 10:21:21.556140 11756 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0930 10:21:21.556166 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0930 10:21:21.950013 11756 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0930 10:21:21.950036 11756 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0930 10:21:22.149983 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0930 10:21:22.267672 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-qxcw9" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:23.154948 11756 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0930 10:21:23.155088 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:23.182168 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:23.967250 11756 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0930 10:21:24.251183 11756 addons.go:234] Setting addon gcp-auth=true in "addons-485025"
I0930 10:21:24.251263 11756 host.go:66] Checking if "addons-485025" exists ...
I0930 10:21:24.252873 11756 cli_runner.go:164] Run: docker container inspect addons-485025 --format={{.State.Status}}
I0930 10:21:24.280611 11756 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0930 10:21:24.280671 11756 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-485025
I0930 10:21:24.297658 11756 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19734-3685/.minikube/machines/addons-485025/id_rsa Username:docker}
I0930 10:21:24.758471 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-qxcw9" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:25.461056 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (8.698408037s)
I0930 10:21:25.461099 11756 addons.go:475] Verifying addon ingress=true in "addons-485025"
I0930 10:21:25.461272 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (8.610004912s)
I0930 10:21:25.461331 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (8.599739493s)
I0930 10:21:25.461177 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (8.69651338s)
I0930 10:21:25.461207 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (8.611021267s)
I0930 10:21:25.463439 11756 out.go:177] * Verifying ingress addon...
I0930 10:21:25.467489 11756 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
W0930 10:21:25.470700 11756 out.go:270] ! Enabling 'default-storageclass' returned an error: running callbacks: [Error making standard the default storage class: Error while marking storage class local-path as non-default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
I0930 10:21:25.471964 11756 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0930 10:21:25.471985 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:25.972182 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:26.474483 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:27.062244 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:27.254092 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-qxcw9" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:27.473674 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:27.971384 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:28.257250 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.298459614s)
I0930 10:21:28.257402 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (11.184384551s)
I0930 10:21:28.257435 11756 addons.go:475] Verifying addon registry=true in "addons-485025"
I0930 10:21:28.257675 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (10.201171881s)
I0930 10:21:28.257772 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (10.000472239s)
I0930 10:21:28.257982 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (9.492778905s)
W0930 10:21:28.258014 11756 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0930 10:21:28.258031 11756 addons.go:475] Verifying addon metrics-server=true in "addons-485025"
I0930 10:21:28.258034 11756 retry.go:31] will retry after 140.689133ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0930 10:21:28.258163 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-configmap.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (7.695934948s)
I0930 10:21:28.258964 11756 out.go:177] * Verifying registry addon...
I0930 10:21:28.259734 11756 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-485025 service yakd-dashboard -n yakd-dashboard
I0930 10:21:28.261464 11756 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0930 10:21:28.264480 11756 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0930 10:21:28.264502 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:28.399212 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0930 10:21:28.550088 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:28.767098 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:29.050421 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:29.150883 11756 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (4.870247124s)
I0930 10:21:29.150799 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (7.000764788s)
I0930 10:21:29.151121 11756 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-485025"
I0930 10:21:29.152580 11756 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0930 10:21:29.152682 11756 out.go:177] * Verifying csi-hostpath-driver addon...
I0930 10:21:29.154390 11756 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0930 10:21:29.155412 11756 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0930 10:21:29.155846 11756 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0930 10:21:29.155869 11756 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0930 10:21:29.162273 11756 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0930 10:21:29.162298 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:29.256658 11756 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0930 10:21:29.256696 11756 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0930 10:21:29.265406 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:29.281269 11756 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0930 10:21:29.281289 11756 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0930 10:21:29.551360 11756 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0930 10:21:29.552282 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:29.660487 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:29.678965 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-qxcw9" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:29.765779 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:29.972072 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:30.161901 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:30.265251 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:30.474119 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:30.661254 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:30.675060 11756 pod_ready.go:98] error getting pod "coredns-7c65d6cfc9-qxcw9" in "kube-system" namespace (skipping!): pods "coredns-7c65d6cfc9-qxcw9" not found
I0930 10:21:30.675089 11756 pod_ready.go:82] duration metric: took 12.502049818s for pod "coredns-7c65d6cfc9-qxcw9" in "kube-system" namespace to be "Ready" ...
E0930 10:21:30.675101 11756 pod_ready.go:67] WaitExtra: waitPodCondition: error getting pod "coredns-7c65d6cfc9-qxcw9" in "kube-system" namespace (skipping!): pods "coredns-7c65d6cfc9-qxcw9" not found
I0930 10:21:30.675112 11756 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace to be "Ready" ...
I0930 10:21:30.765443 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:30.953009 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.553752045s)
I0930 10:21:30.975581 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:30.979753 11756 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.428345524s)
I0930 10:21:30.982008 11756 addons.go:475] Verifying addon gcp-auth=true in "addons-485025"
I0930 10:21:30.985105 11756 out.go:177] * Verifying gcp-auth addon...
I0930 10:21:30.987192 11756 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0930 10:21:31.075163 11756 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0930 10:21:31.160020 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:31.265321 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:31.471817 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:31.659260 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:31.765247 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:31.971255 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:32.173518 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:32.273120 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:32.471688 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:32.660429 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:32.681450 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:32.765804 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:32.971252 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:33.162033 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:33.265064 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:33.471821 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:33.660370 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:33.765521 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:33.972052 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:34.159398 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:34.265376 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:34.471627 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:34.659442 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:34.765346 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:34.972079 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:35.159616 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:35.180466 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:35.265120 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:35.471269 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:35.659702 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:35.766832 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:35.972908 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:36.160268 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:36.264957 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:36.471364 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:36.660353 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:36.765637 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:36.971546 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:37.160096 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:37.265141 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:37.471305 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:37.660262 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:37.681258 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:37.765314 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:37.972074 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:38.160196 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:38.264832 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:38.472157 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:38.659972 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:38.764925 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:38.971425 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:39.159791 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:39.265679 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:39.472080 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:39.659760 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:39.765620 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:40.053567 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:40.295107 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:40.295619 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:40.297050 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:40.471292 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:40.659771 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:40.766619 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:40.971614 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:41.159390 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:41.264729 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:41.472405 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:41.660298 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:41.765592 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:41.972403 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:42.160020 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:42.264711 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:42.472451 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:42.660424 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:42.680056 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:42.765124 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:43.024211 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:43.159336 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:43.265524 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:43.471873 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:43.662094 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:43.765621 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:43.971636 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:44.160737 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:44.264827 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:44.471409 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:44.660421 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:44.680630 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:44.765438 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:44.971864 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:45.159873 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:45.264169 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:45.471122 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:45.660682 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:45.765509 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:45.971946 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:46.160387 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:46.265577 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:46.472054 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:46.660157 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:46.681205 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:46.765253 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:46.971820 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:47.160742 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:47.264419 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0930 10:21:47.471562 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:47.660357 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:47.765913 11756 kapi.go:107] duration metric: took 19.504444181s to wait for kubernetes.io/minikube-addons=registry ...
I0930 10:21:47.972616 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:48.160238 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:48.472084 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:48.660875 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:48.972525 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:49.160495 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:49.179986 11756 pod_ready.go:103] pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace has status "Ready":"False"
I0930 10:21:49.578502 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:49.680065 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:49.971580 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:50.160474 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:50.180548 11756 pod_ready.go:93] pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace has status "Ready":"True"
I0930 10:21:50.180572 11756 pod_ready.go:82] duration metric: took 19.50545169s for pod "coredns-7c65d6cfc9-vdjlp" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.180584 11756 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-485025" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.184893 11756 pod_ready.go:93] pod "etcd-addons-485025" in "kube-system" namespace has status "Ready":"True"
I0930 10:21:50.184916 11756 pod_ready.go:82] duration metric: took 4.322019ms for pod "etcd-addons-485025" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.184928 11756 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-485025" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.189262 11756 pod_ready.go:93] pod "kube-apiserver-addons-485025" in "kube-system" namespace has status "Ready":"True"
I0930 10:21:50.189284 11756 pod_ready.go:82] duration metric: took 4.347746ms for pod "kube-apiserver-addons-485025" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.189295 11756 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-485025" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.193205 11756 pod_ready.go:93] pod "kube-controller-manager-addons-485025" in "kube-system" namespace has status "Ready":"True"
I0930 10:21:50.193223 11756 pod_ready.go:82] duration metric: took 3.920446ms for pod "kube-controller-manager-addons-485025" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.193233 11756 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-r4dfl" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.197200 11756 pod_ready.go:93] pod "kube-proxy-r4dfl" in "kube-system" namespace has status "Ready":"True"
I0930 10:21:50.197220 11756 pod_ready.go:82] duration metric: took 3.980012ms for pod "kube-proxy-r4dfl" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.197231 11756 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-485025" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.471444 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:50.578162 11756 pod_ready.go:93] pod "kube-scheduler-addons-485025" in "kube-system" namespace has status "Ready":"True"
I0930 10:21:50.578188 11756 pod_ready.go:82] duration metric: took 380.947977ms for pod "kube-scheduler-addons-485025" in "kube-system" namespace to be "Ready" ...
I0930 10:21:50.578199 11756 pod_ready.go:39] duration metric: took 32.414407889s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0930 10:21:50.578224 11756 api_server.go:52] waiting for apiserver process to appear ...
I0930 10:21:50.578287 11756 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0930 10:21:50.594523 11756 api_server.go:72] duration metric: took 34.570935152s to wait for apiserver process to appear ...
I0930 10:21:50.594558 11756 api_server.go:88] waiting for apiserver healthz status ...
I0930 10:21:50.594583 11756 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0930 10:21:50.598673 11756 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0930 10:21:50.599623 11756 api_server.go:141] control plane version: v1.31.1
I0930 10:21:50.599649 11756 api_server.go:131] duration metric: took 5.083695ms to wait for apiserver health ...
I0930 10:21:50.599658 11756 system_pods.go:43] waiting for kube-system pods to appear ...
I0930 10:21:50.659891 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:50.782645 11756 system_pods.go:59] 17 kube-system pods found
I0930 10:21:50.782674 11756 system_pods.go:61] "coredns-7c65d6cfc9-vdjlp" [8972b887-927f-4352-9193-7055c500efb6] Running
I0930 10:21:50.782684 11756 system_pods.go:61] "csi-hostpath-attacher-0" [310b2d52-6848-45f4-94f7-76b5950ff8c7] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0930 10:21:50.782690 11756 system_pods.go:61] "csi-hostpath-resizer-0" [6f8e1b1a-be75-49ee-9d72-20cbc0eb9056] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0930 10:21:50.782702 11756 system_pods.go:61] "csi-hostpathplugin-sl6b6" [fb1fc1ba-92b9-4e1d-8888-4233bccc7032] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0930 10:21:50.782708 11756 system_pods.go:61] "etcd-addons-485025" [a1e2dffc-229f-4f2a-9fd7-b616efee8e08] Running
I0930 10:21:50.782714 11756 system_pods.go:61] "kube-apiserver-addons-485025" [817b4bf7-c1c1-4252-90c6-61c314af5c76] Running
I0930 10:21:50.782722 11756 system_pods.go:61] "kube-controller-manager-addons-485025" [df8f60bc-4425-4584-95dd-2d4cd904bd83] Running
I0930 10:21:50.782732 11756 system_pods.go:61] "kube-ingress-dns-minikube" [3f251512-e5ef-4716-9e29-0eb39032d93f] Running
I0930 10:21:50.782737 11756 system_pods.go:61] "kube-proxy-r4dfl" [1f81e52d-d695-4f77-81b5-fb2fd1a5a7c6] Running
I0930 10:21:50.782745 11756 system_pods.go:61] "kube-scheduler-addons-485025" [01d104cb-ac63-4542-ac3c-793c7246b186] Running
I0930 10:21:50.782752 11756 system_pods.go:61] "metrics-server-84c5f94fbc-kvtbh" [213313da-61ef-4454-a082-5c64f6fad3d1] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0930 10:21:50.782757 11756 system_pods.go:61] "nvidia-device-plugin-daemonset-5zsrh" [1716d456-7b54-4982-b487-8bf11f302e7f] Running
I0930 10:21:50.782764 11756 system_pods.go:61] "registry-66c9cd494c-9bg4w" [da79db35-9dbe-40b6-bc10-153757b8bf2a] Running
I0930 10:21:50.782767 11756 system_pods.go:61] "registry-proxy-8lrkc" [0863352b-681f-45ef-a925-ee3ba3eb1198] Running
I0930 10:21:50.782773 11756 system_pods.go:61] "snapshot-controller-56fcc65765-224qv" [a09a0400-f6a3-4cda-953a-7e5738bdf97f] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0930 10:21:50.782781 11756 system_pods.go:61] "snapshot-controller-56fcc65765-gqqd4" [5759f1d0-838c-4956-8358-dce84db71f65] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0930 10:21:50.782787 11756 system_pods.go:61] "storage-provisioner" [1ac3769e-d2e9-40d1-8c38-c2494ba4e962] Running
I0930 10:21:50.782795 11756 system_pods.go:74] duration metric: took 183.129605ms to wait for pod list to return data ...
I0930 10:21:50.782806 11756 default_sa.go:34] waiting for default service account to be created ...
I0930 10:21:50.972220 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:50.977785 11756 default_sa.go:45] found service account: "default"
I0930 10:21:50.977814 11756 default_sa.go:55] duration metric: took 194.998219ms for default service account to be created ...
I0930 10:21:50.977824 11756 system_pods.go:116] waiting for k8s-apps to be running ...
I0930 10:21:51.160181 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:51.256921 11756 system_pods.go:86] 17 kube-system pods found
I0930 10:21:51.256959 11756 system_pods.go:89] "coredns-7c65d6cfc9-vdjlp" [8972b887-927f-4352-9193-7055c500efb6] Running
I0930 10:21:51.256975 11756 system_pods.go:89] "csi-hostpath-attacher-0" [310b2d52-6848-45f4-94f7-76b5950ff8c7] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0930 10:21:51.256986 11756 system_pods.go:89] "csi-hostpath-resizer-0" [6f8e1b1a-be75-49ee-9d72-20cbc0eb9056] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0930 10:21:51.256997 11756 system_pods.go:89] "csi-hostpathplugin-sl6b6" [fb1fc1ba-92b9-4e1d-8888-4233bccc7032] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0930 10:21:51.257007 11756 system_pods.go:89] "etcd-addons-485025" [a1e2dffc-229f-4f2a-9fd7-b616efee8e08] Running
I0930 10:21:51.257015 11756 system_pods.go:89] "kube-apiserver-addons-485025" [817b4bf7-c1c1-4252-90c6-61c314af5c76] Running
I0930 10:21:51.257022 11756 system_pods.go:89] "kube-controller-manager-addons-485025" [df8f60bc-4425-4584-95dd-2d4cd904bd83] Running
I0930 10:21:51.257032 11756 system_pods.go:89] "kube-ingress-dns-minikube" [3f251512-e5ef-4716-9e29-0eb39032d93f] Running
I0930 10:21:51.257037 11756 system_pods.go:89] "kube-proxy-r4dfl" [1f81e52d-d695-4f77-81b5-fb2fd1a5a7c6] Running
I0930 10:21:51.257042 11756 system_pods.go:89] "kube-scheduler-addons-485025" [01d104cb-ac63-4542-ac3c-793c7246b186] Running
I0930 10:21:51.257056 11756 system_pods.go:89] "metrics-server-84c5f94fbc-kvtbh" [213313da-61ef-4454-a082-5c64f6fad3d1] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0930 10:21:51.257070 11756 system_pods.go:89] "nvidia-device-plugin-daemonset-5zsrh" [1716d456-7b54-4982-b487-8bf11f302e7f] Running
I0930 10:21:51.257084 11756 system_pods.go:89] "registry-66c9cd494c-9bg4w" [da79db35-9dbe-40b6-bc10-153757b8bf2a] Running
I0930 10:21:51.257089 11756 system_pods.go:89] "registry-proxy-8lrkc" [0863352b-681f-45ef-a925-ee3ba3eb1198] Running
I0930 10:21:51.257103 11756 system_pods.go:89] "snapshot-controller-56fcc65765-224qv" [a09a0400-f6a3-4cda-953a-7e5738bdf97f] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0930 10:21:51.257112 11756 system_pods.go:89] "snapshot-controller-56fcc65765-gqqd4" [5759f1d0-838c-4956-8358-dce84db71f65] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0930 10:21:51.257125 11756 system_pods.go:89] "storage-provisioner" [1ac3769e-d2e9-40d1-8c38-c2494ba4e962] Running
I0930 10:21:51.257135 11756 system_pods.go:126] duration metric: took 279.304553ms to wait for k8s-apps to be running ...
I0930 10:21:51.257149 11756 system_svc.go:44] waiting for kubelet service to be running ....
I0930 10:21:51.257198 11756 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0930 10:21:51.349033 11756 system_svc.go:56] duration metric: took 91.866671ms WaitForService to wait for kubelet
I0930 10:21:51.349069 11756 kubeadm.go:582] duration metric: took 35.325485832s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0930 10:21:51.349092 11756 node_conditions.go:102] verifying NodePressure condition ...
I0930 10:21:51.379561 11756 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0930 10:21:51.379587 11756 node_conditions.go:123] node cpu capacity is 8
I0930 10:21:51.379599 11756 node_conditions.go:105] duration metric: took 30.502511ms to run NodePressure ...
I0930 10:21:51.379610 11756 start.go:241] waiting for startup goroutines ...
I0930 10:21:51.472446 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:51.659891 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:51.972297 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:52.159997 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:52.471462 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:52.660696 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:52.972077 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:53.159721 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:53.470863 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:53.658953 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:53.972179 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:54.159005 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:54.472208 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:54.660195 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:55.053668 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:55.159681 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:55.471836 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:55.660882 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:55.973638 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:56.160706 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:56.482830 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:56.684421 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:56.972421 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:57.160909 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:57.471399 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:57.660476 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:57.972270 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:58.159825 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:58.472034 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:58.659813 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:58.972635 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:59.160102 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:59.472243 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:21:59.660085 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:21:59.971959 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:00.160451 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:00.471556 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:00.660087 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:00.971444 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:01.253233 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:01.471994 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:01.660395 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:01.972550 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:02.160348 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:02.471020 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:02.673169 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:02.990065 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:03.159829 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:03.484014 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:03.686507 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:03.971609 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:04.160076 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:04.471715 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:04.673558 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:04.972673 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:05.160372 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:05.472200 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:05.660105 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:05.972157 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:06.160414 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:06.471887 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:06.660686 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:06.971767 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:07.160779 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:07.471232 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:07.660463 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:07.972152 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:08.160374 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:08.472312 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:08.659442 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:08.972385 11756 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0930 10:22:09.160547 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:09.472257 11756 kapi.go:107] duration metric: took 44.004768698s to wait for app.kubernetes.io/name=ingress-nginx ...
I0930 10:22:09.660300 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:10.193545 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:10.661983 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:11.159374 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:11.659991 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:12.167241 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:12.660924 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:13.160685 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:13.660007 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:14.160648 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0930 10:22:14.660098 11756 kapi.go:107] duration metric: took 45.504684757s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0930 10:22:54.490077 11756 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0930 10:22:54.490095 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:54.990471 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:55.490239 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:55.990448 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:56.490723 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:56.990789 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:57.490734 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:57.991056 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:58.491240 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:58.989879 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:59.490717 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:22:59.990755 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:00.490456 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:00.990631 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:01.490507 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:01.991194 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:02.490197 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:02.991280 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:03.490853 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:03.990785 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:04.490916 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:04.990874 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:05.490680 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:05.990426 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:06.490540 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:06.990331 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:07.490439 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:07.990223 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:08.490619 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:08.990437 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:09.490084 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:09.992284 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:10.490124 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:10.989670 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:11.490619 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:11.990846 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:12.490921 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:12.991319 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:13.490194 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:13.990034 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:14.490773 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:14.990931 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:15.490773 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:15.990539 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:16.490927 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:16.991088 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:17.489955 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:17.990840 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:18.491023 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:18.989849 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:19.490611 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:19.990609 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:20.490301 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:20.990361 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:21.490433 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:21.990401 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:22.490372 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:22.989989 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:23.489972 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:23.990880 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:24.491181 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:24.990051 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:25.489736 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:25.991011 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:26.490159 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:26.990126 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:27.490421 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:27.990248 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:28.490656 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:28.990547 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:29.490267 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:29.990371 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:30.490164 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:30.989901 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:31.490892 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:31.990731 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:32.490967 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:32.990571 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:33.491003 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:33.990803 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:34.489961 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:34.991332 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:35.489923 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:35.990729 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:36.490751 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:36.990331 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:37.490281 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:37.991055 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:38.490032 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:38.990534 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:39.490261 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:39.990431 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:40.490388 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:40.990498 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:41.490255 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:41.990322 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:42.490406 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:42.990256 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:43.490828 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:43.990669 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:44.490709 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:44.990307 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:45.489820 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:45.990644 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:46.490979 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:46.991353 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:47.490297 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:47.990178 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:48.490475 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:48.990342 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:49.490087 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:49.991445 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:50.490210 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:50.989999 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:51.490913 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:51.990971 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:52.491040 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:52.990064 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:53.490183 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:53.990639 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:54.490553 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:54.990483 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:55.490379 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:55.990378 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:56.490638 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:56.990897 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:57.491344 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:57.990288 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:58.490718 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:58.990559 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:59.490570 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:23:59.990699 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:24:00.490625 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:24:00.990816 11756 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0930 10:24:01.490756 11756 kapi.go:107] duration metric: took 2m30.503563452s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0930 10:24:01.492259 11756 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-485025 cluster.
I0930 10:24:01.493370 11756 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0930 10:24:01.494567 11756 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0930 10:24:01.495736 11756 out.go:177] * Enabled addons: storage-provisioner, ingress-dns, cloud-spanner, nvidia-device-plugin, storage-provisioner-rancher, volcano, metrics-server, inspektor-gadget, yakd, volumesnapshots, registry, ingress, csi-hostpath-driver, gcp-auth
I0930 10:24:01.496938 11756 addons.go:510] duration metric: took 2m45.473326997s for enable addons: enabled=[storage-provisioner ingress-dns cloud-spanner nvidia-device-plugin storage-provisioner-rancher volcano metrics-server inspektor-gadget yakd volumesnapshots registry ingress csi-hostpath-driver gcp-auth]
I0930 10:24:01.496973 11756 start.go:246] waiting for cluster config update ...
I0930 10:24:01.496990 11756 start.go:255] writing updated cluster config ...
I0930 10:24:01.497225 11756 ssh_runner.go:195] Run: rm -f paused
I0930 10:24:01.543722 11756 start.go:600] kubectl: 1.31.1, cluster: 1.31.1 (minor skew: 0)
I0930 10:24:01.545380 11756 out.go:177] * Done! kubectl is now configured to use "addons-485025" cluster and "default" namespace by default
==> Docker <==
Sep 30 10:33:32 addons-485025 dockerd[1345]: time="2024-09-30T10:33:32.452378827Z" level=info msg="ignoring event" container=1b847ea644f063270b61d2b01de8a0822c42ea258b90b3e00508d11616be871c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:32 addons-485025 dockerd[1345]: time="2024-09-30T10:33:32.452439449Z" level=info msg="ignoring event" container=08b845c9bb5e782cca300e8cb53c56dba479de56f0881b2a12abfb3ab2be53ee module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:32 addons-485025 dockerd[1345]: time="2024-09-30T10:33:32.455424469Z" level=info msg="ignoring event" container=bbbc6abfed283ac036674e0ba0dabb86b03835b5b0cf7dd5b1eea9e4ca1e53c3 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:32 addons-485025 dockerd[1345]: time="2024-09-30T10:33:32.456738501Z" level=info msg="ignoring event" container=dbdcf17f6890e5f274d40cc8c1c6b3501f319b9d693ec825a9e0a6fefa0b9d6c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:32 addons-485025 dockerd[1345]: time="2024-09-30T10:33:32.458998200Z" level=info msg="ignoring event" container=e77a76f5c0314c4afcd9bcf737532a2d8c45951531305d1e4a36967b547f60f0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:32 addons-485025 dockerd[1345]: time="2024-09-30T10:33:32.469923687Z" level=info msg="ignoring event" container=9537e1420f35d694bf58609533cb232bcf785f15c591d961bdfbaae1f931572d module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:32 addons-485025 dockerd[1345]: time="2024-09-30T10:33:32.471559377Z" level=info msg="ignoring event" container=437a82de85bd04a61d87ae273433b408faa56a088bba947c29046b7926c2f6e0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:32 addons-485025 dockerd[1345]: time="2024-09-30T10:33:32.683089685Z" level=info msg="ignoring event" container=cc2bb30faf9d28752f9197fd587678bad506cd4502b52e778e4d98711ccef73b module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:32 addons-485025 dockerd[1345]: time="2024-09-30T10:33:32.751674538Z" level=info msg="ignoring event" container=d3256261bb30beba08777f50b27c12450f0361712b3cc22c2972f27a11770d56 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:32 addons-485025 dockerd[1345]: time="2024-09-30T10:33:32.784303491Z" level=info msg="ignoring event" container=7d5373a54a10c9790233c2aecbdb5b5cfcd4d8e7611dd4e4ff0a60bff5f37749 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:35 addons-485025 dockerd[1345]: time="2024-09-30T10:33:35.706425835Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed" spanID=a2b6b40366ec7988 traceID=221e631d16b91bb5c2236e1df0a780b3
Sep 30 10:33:35 addons-485025 dockerd[1345]: time="2024-09-30T10:33:35.708172138Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed" spanID=a2b6b40366ec7988 traceID=221e631d16b91bb5c2236e1df0a780b3
Sep 30 10:33:38 addons-485025 dockerd[1345]: time="2024-09-30T10:33:38.698458485Z" level=info msg="ignoring event" container=25a976feb551798a47539ed87a5163bfdf47ca96d484466c768a705e4123f634 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:38 addons-485025 dockerd[1345]: time="2024-09-30T10:33:38.700800438Z" level=info msg="ignoring event" container=7e305b4df57f8fdedc7d7ce55ab88afc5755c2ee4863684228045acf5a54883a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:38 addons-485025 dockerd[1345]: time="2024-09-30T10:33:38.869967424Z" level=info msg="ignoring event" container=f179ca433309ac874fb839c555a04429bb5c9bea35e34f7c46fbf70eee8712b7 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:38 addons-485025 dockerd[1345]: time="2024-09-30T10:33:38.908388403Z" level=info msg="ignoring event" container=eb251aef0e0c878579c9f96034cc3720e52f37af516796d1ba2c945c59481fb6 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:43 addons-485025 dockerd[1345]: time="2024-09-30T10:33:43.777462129Z" level=info msg="Container failed to exit within 30s of signal 15 - using the force" container=10a2f56d9b7c12ff7d5eed873bf0b40d45aad54fc1200d53444ace0ddfc77f52 spanID=f4b1136ecb71efe3 traceID=773cbbac141462d5c43364849e6cb288
Sep 30 10:33:43 addons-485025 dockerd[1345]: time="2024-09-30T10:33:43.799059936Z" level=info msg="ignoring event" container=10a2f56d9b7c12ff7d5eed873bf0b40d45aad54fc1200d53444ace0ddfc77f52 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:43 addons-485025 dockerd[1345]: time="2024-09-30T10:33:43.906321090Z" level=info msg="ignoring event" container=54e1b7f4f3479773db14b057d84bbeeb4e9a34019d16f6acb4c9156f38bb4239 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:53 addons-485025 dockerd[1345]: time="2024-09-30T10:33:53.225147415Z" level=info msg="ignoring event" container=0dcf2f146ba1673fb88da6a81e12bb5d25aa44a2607fc4dee7305ec5af936194 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:53 addons-485025 dockerd[1345]: time="2024-09-30T10:33:53.672664806Z" level=info msg="ignoring event" container=1a3722819018fefe7024fda3df77dbcf4e0eb732b32ea7395d99c42af465f404 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:53 addons-485025 dockerd[1345]: time="2024-09-30T10:33:53.757185994Z" level=info msg="ignoring event" container=8d55d3c5ad40505cc0f45fcf2bd9701fab556cc9c36c84e605e99195bcbbaf16 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:53 addons-485025 dockerd[1345]: time="2024-09-30T10:33:53.814491395Z" level=info msg="ignoring event" container=9b03e0046bf4ac46cb83479b2188636b7f1429175182042fa8901d99b5dc797a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 30 10:33:53 addons-485025 cri-dockerd[1610]: time="2024-09-30T10:33:53Z" level=info msg="Failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod \"registry-proxy-8lrkc_kube-system\": unexpected command output Device \"eth0\" does not exist.\n with error: exit status 1"
Sep 30 10:33:53 addons-485025 dockerd[1345]: time="2024-09-30T10:33:53.898844622Z" level=info msg="ignoring event" container=155d1341ce754944df17d703d704ad7a180ae9c3e3b6eafc4ac3901b2caaff43 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
eaa809e8b9733 a416a98b71e22 41 seconds ago Exited helper-pod 0 e2c12e07c9b01 helper-pod-delete-pvc-c9d28883-8cdc-411a-b481-ed6040da0be1
f0f0dd4ecfe74 busybox@sha256:c230832bd3b0be59a6c47ed64294f9ce71e91b327957920b6929a0caa8353140 44 seconds ago Exited busybox 0 6fe0d4de55d7b test-local-path
5bc7baea2148e kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 57 seconds ago Running hello-world-app 0 1b572184a0848 hello-world-app-55bf9c44b4-p6p72
a9bef864587df nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf About a minute ago Running nginx 0 0a188c9ac0bcc nginx
d9dbd70a3058a gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 2dc5a071e38fc gcp-auth-89d5ffd79-q69gw
8d55d3c5ad405 gcr.io/k8s-minikube/kube-registry-proxy@sha256:9fd683b2e47c5fded3410c69f414f05cdee737597569f52854347f889b118982 12 minutes ago Exited registry-proxy 0 155d1341ce754 registry-proxy-8lrkc
570be091a2bb9 6e38f40d628db 12 minutes ago Running storage-provisioner 0 3cc41922d950e storage-provisioner
c670266622d75 c69fa2e9cbf5f 12 minutes ago Running coredns 0 efa3991f5c0e7 coredns-7c65d6cfc9-vdjlp
0bb3495be0d08 60c005f310ff3 12 minutes ago Running kube-proxy 0 c077f4cb6ff10 kube-proxy-r4dfl
bf4659e7f16dc 175ffd71cce3d 12 minutes ago Running kube-controller-manager 0 b48241e26c74b kube-controller-manager-addons-485025
23877de9b8f78 6bab7719df100 12 minutes ago Running kube-apiserver 0 3f65ac73d15d4 kube-apiserver-addons-485025
f8a8bdbbb7e99 9aa1fad941575 12 minutes ago Running kube-scheduler 0 c470d7ad6be8a kube-scheduler-addons-485025
706fd73224678 2e96e5913fc06 12 minutes ago Running etcd 0 118577d9fb65b etcd-addons-485025
==> coredns [c670266622d7] <==
[INFO] 10.244.0.21:49086 - 40113 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.006958734s
[INFO] 10.244.0.21:49086 - 40427 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004765025s
[INFO] 10.244.0.21:46442 - 37361 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004837492s
[INFO] 10.244.0.21:33444 - 16585 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004548336s
[INFO] 10.244.0.21:56303 - 44893 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004596588s
[INFO] 10.244.0.21:49292 - 12702 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003750609s
[INFO] 10.244.0.21:59125 - 8484 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005443689s
[INFO] 10.244.0.21:36417 - 29577 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004987177s
[INFO] 10.244.0.21:57508 - 12942 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005424184s
[INFO] 10.244.0.21:56303 - 64659 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004527402s
[INFO] 10.244.0.21:57508 - 43774 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005032492s
[INFO] 10.244.0.21:33444 - 52060 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005331649s
[INFO] 10.244.0.21:59125 - 41716 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005309611s
[INFO] 10.244.0.21:56303 - 5820 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000076372s
[INFO] 10.244.0.21:46442 - 22397 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00541669s
[INFO] 10.244.0.21:49292 - 61737 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005322693s
[INFO] 10.244.0.21:49086 - 53534 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005078074s
[INFO] 10.244.0.21:59125 - 22127 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000056091s
[INFO] 10.244.0.21:36417 - 13024 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00577033s
[INFO] 10.244.0.21:33444 - 53106 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00008443s
[INFO] 10.244.0.21:49292 - 14625 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000079109s
[INFO] 10.244.0.21:57508 - 21381 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000159451s
[INFO] 10.244.0.21:49086 - 59668 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000180918s
[INFO] 10.244.0.21:46442 - 31578 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000072124s
[INFO] 10.244.0.21:36417 - 53996 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000176833s
==> describe nodes <==
Name: addons-485025
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-485025
kubernetes.io/os=linux
minikube.k8s.io/commit=b68b4b088317c82ffa16da1c47933e77f0f5d128
minikube.k8s.io/name=addons-485025
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_30T10_21_11_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-485025
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 30 Sep 2024 10:21:08 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-485025
AcquireTime: <unset>
RenewTime: Mon, 30 Sep 2024 10:33:46 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 30 Sep 2024 10:33:17 +0000 Mon, 30 Sep 2024 10:21:06 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 30 Sep 2024 10:33:17 +0000 Mon, 30 Sep 2024 10:21:06 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 30 Sep 2024 10:33:17 +0000 Mon, 30 Sep 2024 10:21:06 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 30 Sep 2024 10:33:17 +0000 Mon, 30 Sep 2024 10:21:08 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-485025
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859320Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859320Ki
pods: 110
System Info:
Machine ID: 38ad125b66ea4ec2802e491bc8685941
System UUID: c6ded2f0-b3bc-45ed-ab92-54568dd3b5e7
Boot ID: e8f00f6f-835b-4ab0-acbc-ac28d6990f2c
Kernel Version: 5.15.0-1069-gcp
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.3.1
Kubelet Version: v1.31.1
Kube-Proxy Version: v1.31.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (11 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m14s
default hello-world-app-55bf9c44b4-p6p72 0 (0%) 0 (0%) 0 (0%) 0 (0%) 58s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 66s
gcp-auth gcp-auth-89d5ffd79-q69gw 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
kube-system coredns-7c65d6cfc9-vdjlp 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-485025 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-485025 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-485025 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-r4dfl 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-485025 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeHasSufficientMemory 12m (x8 over 12m) kubelet Node addons-485025 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m (x7 over 12m) kubelet Node addons-485025 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m (x7 over 12m) kubelet Node addons-485025 status is now: NodeHasSufficientPID
Normal Starting 12m kubelet Starting kubelet.
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-485025 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-485025 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-485025 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-485025 event: Registered Node addons-485025 in Controller
==> dmesg <==
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 6a bb 72 55 c4 09 08 06
[Sep30 10:22] IPv4: martian source 10.244.0.1 from 10.244.0.17, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 4a f9 0b 40 51 89 08 06
[ +7.499447] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff ba 94 5e c7 3f 4b 08 06
[ +2.369420] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 6e c3 89 f6 d7 7a 08 06
[ +0.040035] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 66 c4 e3 6e 26 f0 08 06
[ +0.602015] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff e2 00 55 ff 4b bc 08 06
[ +19.949078] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 1e 36 8f 17 e5 e1 08 06
[Sep30 10:23] IPv4: martian source 10.244.0.1 from 10.244.0.23, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 66 21 be 13 6f ba 08 06
[ +0.062942] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 8e b7 38 6e cf 74 08 06
[Sep30 10:24] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 6a e5 7e d0 1c 02 08 06
[ +0.000427] IPv4: martian source 10.244.0.25 from 10.244.0.3, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff 36 8e fa 03 c7 39 08 06
[Sep30 10:32] IPv4: martian source 10.244.0.29 from 10.244.0.21, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff ba 94 5e c7 3f 4b 08 06
[ +1.552765] IPv4: martian source 10.244.0.21 from 10.244.0.3, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 36 8e fa 03 c7 39 08 06
==> etcd [706fd7322467] <==
{"level":"info","ts":"2024-09-30T10:21:06.771318Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-485025 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-09-30T10:21:06.771322Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-30T10:21:06.771485Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-30T10:21:06.771547Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-30T10:21:06.771581Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-30T10:21:06.772613Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-30T10:21:06.772802Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-30T10:21:06.773747Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-09-30T10:21:06.773879Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-30T10:21:06.776388Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-30T10:21:06.776479Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-30T10:21:06.776523Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"warn","ts":"2024-09-30T10:21:40.293080Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"135.767918ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-30T10:21:40.293162Z","caller":"traceutil/trace.go:171","msg":"trace[520933380] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:997; }","duration":"135.85455ms","start":"2024-09-30T10:21:40.157293Z","end":"2024-09-30T10:21:40.293147Z","steps":["trace[520933380] 'range keys from in-memory index tree' (duration: 135.726151ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-30T10:21:40.293074Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"116.613397ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods/kube-system/coredns-7c65d6cfc9-vdjlp\" ","response":"range_response_count:1 size:5091"}
{"level":"info","ts":"2024-09-30T10:21:40.293248Z","caller":"traceutil/trace.go:171","msg":"trace[1498556743] range","detail":"{range_begin:/registry/pods/kube-system/coredns-7c65d6cfc9-vdjlp; range_end:; response_count:1; response_revision:997; }","duration":"116.788216ms","start":"2024-09-30T10:21:40.176436Z","end":"2024-09-30T10:21:40.293224Z","steps":["trace[1498556743] 'range keys from in-memory index tree' (duration: 116.545061ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-30T10:21:43.020699Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"105.509808ms","expected-duration":"100ms","prefix":"read-only range ","request":"limit:1 serializable:true keys_only:true ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-30T10:21:43.020754Z","caller":"traceutil/trace.go:171","msg":"trace[671667205] range","detail":"{range_begin:; range_end:; response_count:0; response_revision:1010; }","duration":"105.575946ms","start":"2024-09-30T10:21:42.915167Z","end":"2024-09-30T10:21:43.020743Z","steps":["trace[671667205] 'range keys from in-memory index tree' (duration: 105.500742ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-30T10:21:49.575951Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"106.556625ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-30T10:21:49.576020Z","caller":"traceutil/trace.go:171","msg":"trace[1820763279] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:1049; }","duration":"106.634498ms","start":"2024-09-30T10:21:49.469372Z","end":"2024-09-30T10:21:49.576007Z","steps":["trace[1820763279] 'range keys from in-memory index tree' (duration: 106.504838ms)"],"step_count":1}
{"level":"info","ts":"2024-09-30T10:21:59.850169Z","caller":"traceutil/trace.go:171","msg":"trace[968818960] transaction","detail":"{read_only:false; response_revision:1131; number_of_response:1; }","duration":"151.592515ms","start":"2024-09-30T10:21:59.698555Z","end":"2024-09-30T10:21:59.850147Z","steps":["trace[968818960] 'process raft request' (duration: 86.321396ms)","trace[968818960] 'compare' (duration: 65.092095ms)"],"step_count":2}
{"level":"info","ts":"2024-09-30T10:22:09.905313Z","caller":"traceutil/trace.go:171","msg":"trace[794133770] transaction","detail":"{read_only:false; response_revision:1195; number_of_response:1; }","duration":"143.849263ms","start":"2024-09-30T10:22:09.761444Z","end":"2024-09-30T10:22:09.905293Z","steps":["trace[794133770] 'process raft request' (duration: 62.596136ms)","trace[794133770] 'compare' (duration: 81.140452ms)"],"step_count":2}
{"level":"info","ts":"2024-09-30T10:31:06.886668Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1855}
{"level":"info","ts":"2024-09-30T10:31:06.909860Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1855,"took":"22.700701ms","hash":1581934560,"current-db-size-bytes":8851456,"current-db-size":"8.9 MB","current-db-size-in-use-bytes":4718592,"current-db-size-in-use":"4.7 MB"}
{"level":"info","ts":"2024-09-30T10:31:06.909900Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":1581934560,"revision":1855,"compact-revision":-1}
==> gcp-auth [d9dbd70a3058] <==
2024/09/30 10:24:40 Ready to write response ...
2024/09/30 10:24:40 Ready to marshal response ...
2024/09/30 10:24:40 Ready to write response ...
2024/09/30 10:32:43 Ready to marshal response ...
2024/09/30 10:32:43 Ready to write response ...
2024/09/30 10:32:43 Ready to marshal response ...
2024/09/30 10:32:43 Ready to write response ...
2024/09/30 10:32:43 Ready to marshal response ...
2024/09/30 10:32:43 Ready to write response ...
2024/09/30 10:32:48 Ready to marshal response ...
2024/09/30 10:32:48 Ready to write response ...
2024/09/30 10:32:53 Ready to marshal response ...
2024/09/30 10:32:53 Ready to write response ...
2024/09/30 10:32:56 Ready to marshal response ...
2024/09/30 10:32:56 Ready to write response ...
2024/09/30 10:32:59 Ready to marshal response ...
2024/09/30 10:32:59 Ready to write response ...
2024/09/30 10:33:05 Ready to marshal response ...
2024/09/30 10:33:05 Ready to write response ...
2024/09/30 10:33:05 Ready to marshal response ...
2024/09/30 10:33:05 Ready to write response ...
2024/09/30 10:33:13 Ready to marshal response ...
2024/09/30 10:33:13 Ready to write response ...
2024/09/30 10:33:23 Ready to marshal response ...
2024/09/30 10:33:23 Ready to write response ...
==> kernel <==
10:33:54 up 16 min, 0 users, load average: 1.61, 0.75, 0.46
Linux addons-485025 5.15.0-1069-gcp #77~20.04.1-Ubuntu SMP Sun Sep 1 19:39:16 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [23877de9b8f7] <==
E0930 10:24:32.276233 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"volcano-controllers\" not found]"
W0930 10:24:32.666986 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0930 10:24:32.858580 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0930 10:32:43.583572 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.103.247.206"}
I0930 10:32:48.753738 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0930 10:32:48.916265 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.103.159.233"}
I0930 10:32:49.387575 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0930 10:32:50.399749 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0930 10:32:53.786659 1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
I0930 10:32:56.380679 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.111.56.109"}
I0930 10:33:07.951609 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
E0930 10:33:29.231893 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
I0930 10:33:38.588741 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0930 10:33:38.588796 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0930 10:33:38.600709 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0930 10:33:38.600760 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0930 10:33:38.601540 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0930 10:33:38.601581 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0930 10:33:38.614558 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0930 10:33:38.614613 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0930 10:33:38.653531 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0930 10:33:38.653573 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0930 10:33:39.602270 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W0930 10:33:39.654595 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
W0930 10:33:39.755779 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
==> kube-controller-manager [bf4659e7f16d] <==
W0930 10:33:40.663358 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0930 10:33:40.663399 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0930 10:33:41.232974 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0930 10:33:41.233011 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0930 10:33:43.209306 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0930 10:33:43.209350 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0930 10:33:43.211178 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0930 10:33:43.211207 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0930 10:33:44.209923 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0930 10:33:44.209960 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0930 10:33:45.471307 1 shared_informer.go:313] Waiting for caches to sync for resource quota
I0930 10:33:45.471346 1 shared_informer.go:320] Caches are synced for resource quota
I0930 10:33:45.964465 1 shared_informer.go:313] Waiting for caches to sync for garbage collector
I0930 10:33:45.964525 1 shared_informer.go:320] Caches are synced for garbage collector
W0930 10:33:46.227538 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0930 10:33:46.227582 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0930 10:33:47.297242 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0930 10:33:47.297289 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0930 10:33:48.108151 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0930 10:33:48.108195 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0930 10:33:48.489354 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0930 10:33:48.489390 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0930 10:33:50.705327 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0930 10:33:50.705373 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0930 10:33:53.641436 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="8.017µs"
==> kube-proxy [0bb3495be0d0] <==
I0930 10:21:17.962370 1 server_linux.go:66] "Using iptables proxy"
I0930 10:21:18.462662 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0930 10:21:18.462763 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0930 10:21:18.861527 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0930 10:21:18.861595 1 server_linux.go:169] "Using iptables Proxier"
I0930 10:21:18.865613 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0930 10:21:18.865984 1 server.go:483] "Version info" version="v1.31.1"
I0930 10:21:18.866006 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0930 10:21:18.954475 1 config.go:199] "Starting service config controller"
I0930 10:21:18.954525 1 shared_informer.go:313] Waiting for caches to sync for service config
I0930 10:21:18.954477 1 config.go:105] "Starting endpoint slice config controller"
I0930 10:21:18.954571 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0930 10:21:18.956071 1 config.go:328] "Starting node config controller"
I0930 10:21:18.956083 1 shared_informer.go:313] Waiting for caches to sync for node config
I0930 10:21:19.055344 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0930 10:21:19.055423 1 shared_informer.go:320] Caches are synced for service config
I0930 10:21:19.056282 1 shared_informer.go:320] Caches are synced for node config
==> kube-scheduler [f8a8bdbbb7e9] <==
E0930 10:21:08.352928 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
E0930 10:21:08.352939 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0930 10:21:08.353032 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
W0930 10:21:08.353126 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0930 10:21:08.353149 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0930 10:21:08.353150 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0930 10:21:08.353160 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
W0930 10:21:08.353061 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0930 10:21:08.353179 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0930 10:21:08.353186 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0930 10:21:09.180801 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0930 10:21:09.180844 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0930 10:21:09.245392 1 reflector.go:561] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0930 10:21:09.245431 1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
W0930 10:21:09.250626 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0930 10:21:09.250659 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0930 10:21:09.272720 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0930 10:21:09.272768 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0930 10:21:09.301983 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0930 10:21:09.302030 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0930 10:21:09.398345 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0930 10:21:09.398381 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0930 10:21:09.405589 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0930 10:21:09.405645 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
I0930 10:21:12.475293 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 30 10:33:44 addons-485025 kubelet[2439]: I0930 10:33:44.121672 2439 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11ae3e07-cdbe-4782-9ad2-bd52db2fb6dd-kube-api-access-5kfzs" (OuterVolumeSpecName: "kube-api-access-5kfzs") pod "11ae3e07-cdbe-4782-9ad2-bd52db2fb6dd" (UID: "11ae3e07-cdbe-4782-9ad2-bd52db2fb6dd"). InnerVolumeSpecName "kube-api-access-5kfzs". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 30 10:33:44 addons-485025 kubelet[2439]: I0930 10:33:44.220126 2439 reconciler_common.go:288] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/11ae3e07-cdbe-4782-9ad2-bd52db2fb6dd-config-volume\") on node \"addons-485025\" DevicePath \"\""
Sep 30 10:33:44 addons-485025 kubelet[2439]: I0930 10:33:44.220154 2439 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-5kfzs\" (UniqueName: \"kubernetes.io/projected/11ae3e07-cdbe-4782-9ad2-bd52db2fb6dd-kube-api-access-5kfzs\") on node \"addons-485025\" DevicePath \"\""
Sep 30 10:33:44 addons-485025 kubelet[2439]: I0930 10:33:44.736739 2439 scope.go:117] "RemoveContainer" containerID="10a2f56d9b7c12ff7d5eed873bf0b40d45aad54fc1200d53444ace0ddfc77f52"
Sep 30 10:33:44 addons-485025 kubelet[2439]: I0930 10:33:44.750182 2439 scope.go:117] "RemoveContainer" containerID="10a2f56d9b7c12ff7d5eed873bf0b40d45aad54fc1200d53444ace0ddfc77f52"
Sep 30 10:33:44 addons-485025 kubelet[2439]: E0930 10:33:44.750825 2439 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 10a2f56d9b7c12ff7d5eed873bf0b40d45aad54fc1200d53444ace0ddfc77f52" containerID="10a2f56d9b7c12ff7d5eed873bf0b40d45aad54fc1200d53444ace0ddfc77f52"
Sep 30 10:33:44 addons-485025 kubelet[2439]: I0930 10:33:44.750858 2439 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"10a2f56d9b7c12ff7d5eed873bf0b40d45aad54fc1200d53444ace0ddfc77f52"} err="failed to get container status \"10a2f56d9b7c12ff7d5eed873bf0b40d45aad54fc1200d53444ace0ddfc77f52\": rpc error: code = Unknown desc = Error response from daemon: No such container: 10a2f56d9b7c12ff7d5eed873bf0b40d45aad54fc1200d53444ace0ddfc77f52"
Sep 30 10:33:46 addons-485025 kubelet[2439]: I0930 10:33:46.657171 2439 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11ae3e07-cdbe-4782-9ad2-bd52db2fb6dd" path="/var/lib/kubelet/pods/11ae3e07-cdbe-4782-9ad2-bd52db2fb6dd/volumes"
Sep 30 10:33:50 addons-485025 kubelet[2439]: E0930 10:33:50.651818 2439 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox\\\"\"" pod="default/registry-test" podUID="bf96a69d-60dc-47c2-b018-cc9bd3efd4d6"
Sep 30 10:33:50 addons-485025 kubelet[2439]: E0930 10:33:50.651879 2439 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="fae7290f-6bd0-4d2a-ae59-c439a980c2fa"
Sep 30 10:33:53 addons-485025 kubelet[2439]: I0930 10:33:53.375343 2439 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr7cf\" (UniqueName: \"kubernetes.io/projected/bf96a69d-60dc-47c2-b018-cc9bd3efd4d6-kube-api-access-nr7cf\") pod \"bf96a69d-60dc-47c2-b018-cc9bd3efd4d6\" (UID: \"bf96a69d-60dc-47c2-b018-cc9bd3efd4d6\") "
Sep 30 10:33:53 addons-485025 kubelet[2439]: I0930 10:33:53.375401 2439 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/bf96a69d-60dc-47c2-b018-cc9bd3efd4d6-gcp-creds\") pod \"bf96a69d-60dc-47c2-b018-cc9bd3efd4d6\" (UID: \"bf96a69d-60dc-47c2-b018-cc9bd3efd4d6\") "
Sep 30 10:33:53 addons-485025 kubelet[2439]: I0930 10:33:53.375477 2439 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf96a69d-60dc-47c2-b018-cc9bd3efd4d6-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "bf96a69d-60dc-47c2-b018-cc9bd3efd4d6" (UID: "bf96a69d-60dc-47c2-b018-cc9bd3efd4d6"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Sep 30 10:33:53 addons-485025 kubelet[2439]: I0930 10:33:53.377052 2439 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf96a69d-60dc-47c2-b018-cc9bd3efd4d6-kube-api-access-nr7cf" (OuterVolumeSpecName: "kube-api-access-nr7cf") pod "bf96a69d-60dc-47c2-b018-cc9bd3efd4d6" (UID: "bf96a69d-60dc-47c2-b018-cc9bd3efd4d6"). InnerVolumeSpecName "kube-api-access-nr7cf". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 30 10:33:53 addons-485025 kubelet[2439]: I0930 10:33:53.476280 2439 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-nr7cf\" (UniqueName: \"kubernetes.io/projected/bf96a69d-60dc-47c2-b018-cc9bd3efd4d6-kube-api-access-nr7cf\") on node \"addons-485025\" DevicePath \"\""
Sep 30 10:33:53 addons-485025 kubelet[2439]: I0930 10:33:53.476317 2439 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/bf96a69d-60dc-47c2-b018-cc9bd3efd4d6-gcp-creds\") on node \"addons-485025\" DevicePath \"\""
Sep 30 10:33:53 addons-485025 kubelet[2439]: I0930 10:33:53.897339 2439 scope.go:117] "RemoveContainer" containerID="1a3722819018fefe7024fda3df77dbcf4e0eb732b32ea7395d99c42af465f404"
Sep 30 10:33:53 addons-485025 kubelet[2439]: I0930 10:33:53.979269 2439 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rw65\" (UniqueName: \"kubernetes.io/projected/da79db35-9dbe-40b6-bc10-153757b8bf2a-kube-api-access-7rw65\") pod \"da79db35-9dbe-40b6-bc10-153757b8bf2a\" (UID: \"da79db35-9dbe-40b6-bc10-153757b8bf2a\") "
Sep 30 10:33:53 addons-485025 kubelet[2439]: I0930 10:33:53.981161 2439 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da79db35-9dbe-40b6-bc10-153757b8bf2a-kube-api-access-7rw65" (OuterVolumeSpecName: "kube-api-access-7rw65") pod "da79db35-9dbe-40b6-bc10-153757b8bf2a" (UID: "da79db35-9dbe-40b6-bc10-153757b8bf2a"). InnerVolumeSpecName "kube-api-access-7rw65". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 30 10:33:54 addons-485025 kubelet[2439]: I0930 10:33:54.079729 2439 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfd4n\" (UniqueName: \"kubernetes.io/projected/0863352b-681f-45ef-a925-ee3ba3eb1198-kube-api-access-bfd4n\") pod \"0863352b-681f-45ef-a925-ee3ba3eb1198\" (UID: \"0863352b-681f-45ef-a925-ee3ba3eb1198\") "
Sep 30 10:33:54 addons-485025 kubelet[2439]: I0930 10:33:54.079806 2439 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-7rw65\" (UniqueName: \"kubernetes.io/projected/da79db35-9dbe-40b6-bc10-153757b8bf2a-kube-api-access-7rw65\") on node \"addons-485025\" DevicePath \"\""
Sep 30 10:33:54 addons-485025 kubelet[2439]: I0930 10:33:54.081511 2439 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0863352b-681f-45ef-a925-ee3ba3eb1198-kube-api-access-bfd4n" (OuterVolumeSpecName: "kube-api-access-bfd4n") pod "0863352b-681f-45ef-a925-ee3ba3eb1198" (UID: "0863352b-681f-45ef-a925-ee3ba3eb1198"). InnerVolumeSpecName "kube-api-access-bfd4n". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 30 10:33:54 addons-485025 kubelet[2439]: I0930 10:33:54.179990 2439 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-bfd4n\" (UniqueName: \"kubernetes.io/projected/0863352b-681f-45ef-a925-ee3ba3eb1198-kube-api-access-bfd4n\") on node \"addons-485025\" DevicePath \"\""
Sep 30 10:33:54 addons-485025 kubelet[2439]: I0930 10:33:54.656977 2439 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf96a69d-60dc-47c2-b018-cc9bd3efd4d6" path="/var/lib/kubelet/pods/bf96a69d-60dc-47c2-b018-cc9bd3efd4d6/volumes"
Sep 30 10:33:54 addons-485025 kubelet[2439]: I0930 10:33:54.657305 2439 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da79db35-9dbe-40b6-bc10-153757b8bf2a" path="/var/lib/kubelet/pods/da79db35-9dbe-40b6-bc10-153757b8bf2a/volumes"
==> storage-provisioner [570be091a2bb] <==
I0930 10:21:23.561599 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0930 10:21:23.649937 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0930 10:21:23.649986 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0930 10:21:23.658729 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0930 10:21:23.658996 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-485025_fec087af-ed5a-43de-b9e8-de88339482b5!
I0930 10:21:23.659567 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"dda72a52-478f-4e63-98b9-deeb94aee097", APIVersion:"v1", ResourceVersion:"611", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-485025_fec087af-ed5a-43de-b9e8-de88339482b5 became leader
I0930 10:21:23.759464 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-485025_fec087af-ed5a-43de-b9e8-de88339482b5!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-485025 -n addons-485025
helpers_test.go:261: (dbg) Run: kubectl --context addons-485025 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-485025 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-485025 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-485025/192.168.49.2
Start Time: Mon, 30 Sep 2024 10:24:40 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.27
IPs:
IP: 10.244.0.27
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-fmtkn (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-fmtkn:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m15s default-scheduler Successfully assigned default/busybox to addons-485025
Normal Pulling 7m43s (x4 over 9m14s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m43s (x4 over 9m14s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m43s (x4 over 9m14s) kubelet Error: ErrImagePull
Warning Failed 7m32s (x6 over 9m14s) kubelet Error: ImagePullBackOff
Normal BackOff 4m8s (x21 over 9m14s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (72.32s)