=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:328: registry stabilized in 3.062162ms
addons_test.go:330: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-84svt" [d2e45ba0-4b0a-4648-a233-1dfc5982c286] Running
addons_test.go:330: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.003876396s
addons_test.go:333: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-s7k45" [4fbca207-de93-4adb-baa8-2219f829573b] Running
addons_test.go:333: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 6.006434631s
addons_test.go:338: (dbg) Run: kubectl --context addons-711398 delete po -l run=registry-test --now
addons_test.go:343: (dbg) Run: kubectl --context addons-711398 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:343: (dbg) Non-zero exit: kubectl --context addons-711398 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.132721189s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:345: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-711398 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:349: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:357: (dbg) Run: out/minikube-linux-arm64 -p addons-711398 ip
2024/09/20 19:35:05 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:386: (dbg) Run: out/minikube-linux-arm64 -p addons-711398 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-711398
helpers_test.go:235: (dbg) docker inspect addons-711398:
-- stdout --
[
{
"Id": "5e42a3b36384066d8ca5004aa14f662e66bf1ada0e09552963e5d4c92df8a2e1",
"Created": "2024-09-20T19:21:51.507733754Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 723627,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-20T19:21:51.662805735Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:f8be4f9f9351784955e36c0e64d55ad19451839d9f6d0c057285eb8f9072963b",
"ResolvConfPath": "/var/lib/docker/containers/5e42a3b36384066d8ca5004aa14f662e66bf1ada0e09552963e5d4c92df8a2e1/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/5e42a3b36384066d8ca5004aa14f662e66bf1ada0e09552963e5d4c92df8a2e1/hostname",
"HostsPath": "/var/lib/docker/containers/5e42a3b36384066d8ca5004aa14f662e66bf1ada0e09552963e5d4c92df8a2e1/hosts",
"LogPath": "/var/lib/docker/containers/5e42a3b36384066d8ca5004aa14f662e66bf1ada0e09552963e5d4c92df8a2e1/5e42a3b36384066d8ca5004aa14f662e66bf1ada0e09552963e5d4c92df8a2e1-json.log",
"Name": "/addons-711398",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-711398:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {}
},
"NetworkMode": "addons-711398",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/85859bf9967404372e4a23dd6e47fd89fd64b995bb138837c37719a33c6400cc-init/diff:/var/lib/docker/overlay2/49b3229d349a779acfb3b52fb14a5968187f2ddeb7c959acb87eba75b03cb72a/diff",
"MergedDir": "/var/lib/docker/overlay2/85859bf9967404372e4a23dd6e47fd89fd64b995bb138837c37719a33c6400cc/merged",
"UpperDir": "/var/lib/docker/overlay2/85859bf9967404372e4a23dd6e47fd89fd64b995bb138837c37719a33c6400cc/diff",
"WorkDir": "/var/lib/docker/overlay2/85859bf9967404372e4a23dd6e47fd89fd64b995bb138837c37719a33c6400cc/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-711398",
"Source": "/var/lib/docker/volumes/addons-711398/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-711398",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-711398",
"name.minikube.sigs.k8s.io": "addons-711398",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "7a47e25f1d6ae56c1b56db44ea99fa5ec6dc229d77e2e46ca82cf35f48b64148",
"SandboxKey": "/var/run/docker/netns/7a47e25f1d6a",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-711398": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "714a7f9ad63b2b663211e1dda48960e7c9687032217f0d1bb937afc5ee3d88fa",
"EndpointID": "a04c1d69bc10fdf487757b78b1693c3e171c4f31eb181fa9d42b199c09503f9f",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-711398",
"5e42a3b36384"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-arm64 status --format={{.Host}} -p addons-711398 -n addons-711398
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-arm64 -p addons-711398 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-arm64 -p addons-711398 logs -n 25: (1.216720192s)
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| start | -o=json --download-only | download-only-164565 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | |
| | -p download-only-164565 | | | | | |
| | --force --alsologtostderr | | | | | |
| | --kubernetes-version=v1.20.0 | | | | | |
| | --container-runtime=docker | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | --all | minikube | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | 20 Sep 24 19:21 UTC |
| delete | -p download-only-164565 | download-only-164565 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | 20 Sep 24 19:21 UTC |
| start | -o=json --download-only | download-only-090878 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | |
| | -p download-only-090878 | | | | | |
| | --force --alsologtostderr | | | | | |
| | --kubernetes-version=v1.31.1 | | | | | |
| | --container-runtime=docker | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | --all | minikube | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | 20 Sep 24 19:21 UTC |
| delete | -p download-only-090878 | download-only-090878 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | 20 Sep 24 19:21 UTC |
| delete | -p download-only-164565 | download-only-164565 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | 20 Sep 24 19:21 UTC |
| delete | -p download-only-090878 | download-only-090878 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | 20 Sep 24 19:21 UTC |
| start | --download-only -p | download-docker-499597 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | |
| | download-docker-499597 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p download-docker-499597 | download-docker-499597 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | 20 Sep 24 19:21 UTC |
| start | --download-only -p | binary-mirror-896186 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | |
| | binary-mirror-896186 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:43279 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-896186 | binary-mirror-896186 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | 20 Sep 24 19:21 UTC |
| addons | disable dashboard -p | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | |
| | addons-711398 | | | | | |
| addons | enable dashboard -p | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | |
| | addons-711398 | | | | | |
| start | -p addons-711398 --wait=true | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:21 UTC | 20 Sep 24 19:25 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| addons | addons-711398 addons disable | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:25 UTC | 20 Sep 24 19:25 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | addons-711398 addons disable | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:33 UTC | 20 Sep 24 19:34 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | addons-711398 addons | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:34 UTC | 20 Sep 24 19:34 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-711398 addons | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:34 UTC | 20 Sep 24 19:34 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:34 UTC | 20 Sep 24 19:34 UTC |
| | -p addons-711398 | | | | | |
| ssh | addons-711398 ssh cat | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:34 UTC | 20 Sep 24 19:34 UTC |
| | /opt/local-path-provisioner/pvc-9a9bf7c2-70be-4ebd-8920-1988957db53e_default_test-pvc/file1 | | | | | |
| addons | addons-711398 addons disable | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:34 UTC | |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-711398 ip | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:35 UTC | 20 Sep 24 19:35 UTC |
| addons | addons-711398 addons disable | addons-711398 | jenkins | v1.34.0 | 20 Sep 24 19:35 UTC | 20 Sep 24 19:35 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/20 19:21:27
Running on machine: ip-172-31-29-130
Binary: Built with gc go1.23.0 for linux/arm64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0920 19:21:27.391334 723137 out.go:345] Setting OutFile to fd 1 ...
I0920 19:21:27.391502 723137 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0920 19:21:27.391531 723137 out.go:358] Setting ErrFile to fd 2...
I0920 19:21:27.391552 723137 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0920 19:21:27.391826 723137 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19678-715609/.minikube/bin
I0920 19:21:27.392386 723137 out.go:352] Setting JSON to false
I0920 19:21:27.393284 723137 start.go:129] hostinfo: {"hostname":"ip-172-31-29-130","uptime":11039,"bootTime":1726849049,"procs":151,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1070-aws","kernelArch":"aarch64","virtualizationSystem":"","virtualizationRole":"","hostId":"36adf542-ef4f-4e2d-a0c8-6868d1383ff9"}
I0920 19:21:27.393384 723137 start.go:139] virtualization:
I0920 19:21:27.395096 723137 out.go:177] * [addons-711398] minikube v1.34.0 on Ubuntu 20.04 (arm64)
I0920 19:21:27.396788 723137 out.go:177] - MINIKUBE_LOCATION=19678
I0920 19:21:27.396968 723137 notify.go:220] Checking for updates...
I0920 19:21:27.399623 723137 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0920 19:21:27.400883 723137 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19678-715609/kubeconfig
I0920 19:21:27.402474 723137 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19678-715609/.minikube
I0920 19:21:27.403632 723137 out.go:177] - MINIKUBE_BIN=out/minikube-linux-arm64
I0920 19:21:27.404923 723137 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0920 19:21:27.406440 723137 driver.go:394] Setting default libvirt URI to qemu:///system
I0920 19:21:27.427139 723137 docker.go:123] docker version: linux-27.3.1:Docker Engine - Community
I0920 19:21:27.427282 723137 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0920 19:21:27.486388 723137 info.go:266] docker info: {ID:U5VK:ZNT5:35M3:FHLW:Q7TL:ELFX:BNAG:AV4T:UD2H:SK5L:SEJV:SJJL Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:29 OomKillDisable:true NGoroutines:44 SystemTime:2024-09-20 19:21:27.476652284 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1070-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:aar
ch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214839296 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-29-130 Labels:[] ExperimentalBuild:false ServerVersion:27.3.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErro
rs:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.17.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.7]] Warnings:<nil>}}
I0920 19:21:27.486500 723137 docker.go:318] overlay module found
I0920 19:21:27.487840 723137 out.go:177] * Using the docker driver based on user configuration
I0920 19:21:27.488896 723137 start.go:297] selected driver: docker
I0920 19:21:27.488916 723137 start.go:901] validating driver "docker" against <nil>
I0920 19:21:27.488929 723137 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0920 19:21:27.489608 723137 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0920 19:21:27.542189 723137 info.go:266] docker info: {ID:U5VK:ZNT5:35M3:FHLW:Q7TL:ELFX:BNAG:AV4T:UD2H:SK5L:SEJV:SJJL Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:29 OomKillDisable:true NGoroutines:44 SystemTime:2024-09-20 19:21:27.533021108 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1070-aws OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:aar
ch64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:2 MemTotal:8214839296 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ip-172-31-29-130 Labels:[] ExperimentalBuild:false ServerVersion:27.3.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErro
rs:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.17.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.7]] Warnings:<nil>}}
I0920 19:21:27.542399 723137 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0920 19:21:27.542621 723137 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0920 19:21:27.543871 723137 out.go:177] * Using Docker driver with root privileges
I0920 19:21:27.545180 723137 cni.go:84] Creating CNI manager for ""
I0920 19:21:27.545265 723137 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0920 19:21:27.545280 723137 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0920 19:21:27.545368 723137 start.go:340] cluster config:
{Name:addons-711398 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-711398 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0920 19:21:27.546704 723137 out.go:177] * Starting "addons-711398" primary control-plane node in "addons-711398" cluster
I0920 19:21:27.547732 723137 cache.go:121] Beginning downloading kic base image for docker with docker
I0920 19:21:27.548991 723137 out.go:177] * Pulling base image v0.0.45-1726589491-19662 ...
I0920 19:21:27.550372 723137 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0920 19:21:27.550422 723137 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19678-715609/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-arm64.tar.lz4
I0920 19:21:27.550435 723137 cache.go:56] Caching tarball of preloaded images
I0920 19:21:27.550473 723137 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 in local docker daemon
I0920 19:21:27.550518 723137 preload.go:172] Found /home/jenkins/minikube-integration/19678-715609/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-arm64.tar.lz4 in cache, skipping download
I0920 19:21:27.550528 723137 cache.go:59] Finished verifying existence of preloaded tar for v1.31.1 on docker
I0920 19:21:27.550896 723137 profile.go:143] Saving config to /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/config.json ...
I0920 19:21:27.550964 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/config.json: {Name:mk96f1d698a5d5182bc7f62f1616f96a768bada0 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:27.565053 723137 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 to local cache
I0920 19:21:27.565174 723137 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 in local cache directory
I0920 19:21:27.565197 723137 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 in local cache directory, skipping pull
I0920 19:21:27.565206 723137 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 exists in cache, skipping pull
I0920 19:21:27.565213 723137 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 as a tarball
I0920 19:21:27.565218 723137 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 from local cache
I0920 19:21:44.777918 723137 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 from cached tarball
I0920 19:21:44.777959 723137 cache.go:194] Successfully downloaded all kic artifacts
I0920 19:21:44.777989 723137 start.go:360] acquireMachinesLock for addons-711398: {Name:mk21025134b424beb2eccb2fad371095a8edea53 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0920 19:21:44.778133 723137 start.go:364] duration metric: took 120.818µs to acquireMachinesLock for "addons-711398"
I0920 19:21:44.778164 723137 start.go:93] Provisioning new machine with config: &{Name:addons-711398 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-711398 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0920 19:21:44.778252 723137 start.go:125] createHost starting for "" (driver="docker")
I0920 19:21:44.780053 723137 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0920 19:21:44.780323 723137 start.go:159] libmachine.API.Create for "addons-711398" (driver="docker")
I0920 19:21:44.780365 723137 client.go:168] LocalClient.Create starting
I0920 19:21:44.780552 723137 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19678-715609/.minikube/certs/ca.pem
I0920 19:21:45.066383 723137 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19678-715609/.minikube/certs/cert.pem
I0920 19:21:45.725010 723137 cli_runner.go:164] Run: docker network inspect addons-711398 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0920 19:21:45.741026 723137 cli_runner.go:211] docker network inspect addons-711398 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0920 19:21:45.741125 723137 network_create.go:284] running [docker network inspect addons-711398] to gather additional debugging logs...
I0920 19:21:45.741148 723137 cli_runner.go:164] Run: docker network inspect addons-711398
W0920 19:21:45.758263 723137 cli_runner.go:211] docker network inspect addons-711398 returned with exit code 1
I0920 19:21:45.758300 723137 network_create.go:287] error running [docker network inspect addons-711398]: docker network inspect addons-711398: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-711398 not found
I0920 19:21:45.758315 723137 network_create.go:289] output of [docker network inspect addons-711398]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-711398 not found
** /stderr **
I0920 19:21:45.758438 723137 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0920 19:21:45.780740 723137 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0x4001805d90}
I0920 19:21:45.780792 723137 network_create.go:124] attempt to create docker network addons-711398 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0920 19:21:45.780852 723137 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-711398 addons-711398
I0920 19:21:45.873983 723137 network_create.go:108] docker network addons-711398 192.168.49.0/24 created
I0920 19:21:45.874016 723137 kic.go:121] calculated static IP "192.168.49.2" for the "addons-711398" container
I0920 19:21:45.874092 723137 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0920 19:21:45.889537 723137 cli_runner.go:164] Run: docker volume create addons-711398 --label name.minikube.sigs.k8s.io=addons-711398 --label created_by.minikube.sigs.k8s.io=true
I0920 19:21:45.912992 723137 oci.go:103] Successfully created a docker volume addons-711398
I0920 19:21:45.913102 723137 cli_runner.go:164] Run: docker run --rm --name addons-711398-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-711398 --entrypoint /usr/bin/test -v addons-711398:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -d /var/lib
I0920 19:21:47.780168 723137 cli_runner.go:217] Completed: docker run --rm --name addons-711398-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-711398 --entrypoint /usr/bin/test -v addons-711398:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -d /var/lib: (1.866949845s)
I0920 19:21:47.780202 723137 oci.go:107] Successfully prepared a docker volume addons-711398
I0920 19:21:47.780230 723137 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0920 19:21:47.780252 723137 kic.go:194] Starting extracting preloaded images to volume ...
I0920 19:21:47.780323 723137 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19678-715609/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v addons-711398:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -I lz4 -xf /preloaded.tar -C /extractDir
I0920 19:21:51.443826 723137 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19678-715609/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-arm64.tar.lz4:/preloaded.tar:ro -v addons-711398:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -I lz4 -xf /preloaded.tar -C /extractDir: (3.663461475s)
I0920 19:21:51.443859 723137 kic.go:203] duration metric: took 3.663604553s to extract preloaded images to volume ...
W0920 19:21:51.443999 723137 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0920 19:21:51.444169 723137 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0920 19:21:51.493090 723137 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-711398 --name addons-711398 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-711398 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-711398 --network addons-711398 --ip 192.168.49.2 --volume addons-711398:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4
I0920 19:21:51.833700 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Running}}
I0920 19:21:51.857814 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:21:51.880342 723137 cli_runner.go:164] Run: docker exec addons-711398 stat /var/lib/dpkg/alternatives/iptables
I0920 19:21:51.953735 723137 oci.go:144] the created container "addons-711398" has a running status.
I0920 19:21:51.953768 723137 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa...
I0920 19:21:52.934149 723137 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0920 19:21:52.963453 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:21:52.982108 723137 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0920 19:21:52.982129 723137 kic_runner.go:114] Args: [docker exec --privileged addons-711398 chown docker:docker /home/docker/.ssh/authorized_keys]
I0920 19:21:53.040012 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:21:53.056862 723137 machine.go:93] provisionDockerMachine start ...
I0920 19:21:53.056963 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:53.073636 723137 main.go:141] libmachine: Using SSH client type: native
I0920 19:21:53.073924 723137 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x413650] 0x415e90 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 19:21:53.073940 723137 main.go:141] libmachine: About to run SSH command:
hostname
I0920 19:21:53.219825 723137 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-711398
I0920 19:21:53.219856 723137 ubuntu.go:169] provisioning hostname "addons-711398"
I0920 19:21:53.219950 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:53.240832 723137 main.go:141] libmachine: Using SSH client type: native
I0920 19:21:53.241105 723137 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x413650] 0x415e90 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 19:21:53.241124 723137 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-711398 && echo "addons-711398" | sudo tee /etc/hostname
I0920 19:21:53.397041 723137 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-711398
I0920 19:21:53.397126 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:53.414962 723137 main.go:141] libmachine: Using SSH client type: native
I0920 19:21:53.415209 723137 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x413650] 0x415e90 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 19:21:53.415227 723137 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-711398' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-711398/g' /etc/hosts;
else
echo '127.0.1.1 addons-711398' | sudo tee -a /etc/hosts;
fi
fi
I0920 19:21:53.560348 723137 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0920 19:21:53.560373 723137 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19678-715609/.minikube CaCertPath:/home/jenkins/minikube-integration/19678-715609/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19678-715609/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19678-715609/.minikube}
I0920 19:21:53.560404 723137 ubuntu.go:177] setting up certificates
I0920 19:21:53.560414 723137 provision.go:84] configureAuth start
I0920 19:21:53.560477 723137 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-711398
I0920 19:21:53.578940 723137 provision.go:143] copyHostCerts
I0920 19:21:53.579026 723137 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19678-715609/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19678-715609/.minikube/ca.pem (1078 bytes)
I0920 19:21:53.579140 723137 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19678-715609/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19678-715609/.minikube/cert.pem (1123 bytes)
I0920 19:21:53.579192 723137 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19678-715609/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19678-715609/.minikube/key.pem (1679 bytes)
I0920 19:21:53.579238 723137 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19678-715609/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19678-715609/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19678-715609/.minikube/certs/ca-key.pem org=jenkins.addons-711398 san=[127.0.0.1 192.168.49.2 addons-711398 localhost minikube]
I0920 19:21:53.763714 723137 provision.go:177] copyRemoteCerts
I0920 19:21:53.763794 723137 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0920 19:21:53.763840 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:53.780559 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:21:53.880773 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0920 19:21:53.905369 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0920 19:21:53.929129 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0920 19:21:53.953134 723137 provision.go:87] duration metric: took 392.704135ms to configureAuth
I0920 19:21:53.953161 723137 ubuntu.go:193] setting minikube options for container-runtime
I0920 19:21:53.953376 723137 config.go:182] Loaded profile config "addons-711398": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0920 19:21:53.953442 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:53.970509 723137 main.go:141] libmachine: Using SSH client type: native
I0920 19:21:53.970771 723137 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x413650] 0x415e90 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 19:21:53.970787 723137 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0920 19:21:54.121257 723137 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0920 19:21:54.121279 723137 ubuntu.go:71] root file system type: overlay
I0920 19:21:54.121404 723137 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0920 19:21:54.121471 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:54.138976 723137 main.go:141] libmachine: Using SSH client type: native
I0920 19:21:54.139225 723137 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x413650] 0x415e90 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 19:21:54.139306 723137 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0920 19:21:54.296585 723137 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0920 19:21:54.296672 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:54.313359 723137 main.go:141] libmachine: Using SSH client type: native
I0920 19:21:54.313612 723137 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x413650] 0x415e90 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0920 19:21:54.313636 723137 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0920 19:21:55.135441 723137 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-09-06 12:06:36.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-20 19:21:54.291909157 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0920 19:21:55.135475 723137 machine.go:96] duration metric: took 2.078589043s to provisionDockerMachine
I0920 19:21:55.135492 723137 client.go:171] duration metric: took 10.355116446s to LocalClient.Create
I0920 19:21:55.135508 723137 start.go:167] duration metric: took 10.355185933s to libmachine.API.Create "addons-711398"
I0920 19:21:55.135520 723137 start.go:293] postStartSetup for "addons-711398" (driver="docker")
I0920 19:21:55.135533 723137 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0920 19:21:55.135605 723137 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0920 19:21:55.135651 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:55.154141 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:21:55.258872 723137 ssh_runner.go:195] Run: cat /etc/os-release
I0920 19:21:55.262886 723137 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0920 19:21:55.262925 723137 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0920 19:21:55.262940 723137 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0920 19:21:55.262953 723137 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0920 19:21:55.262968 723137 filesync.go:126] Scanning /home/jenkins/minikube-integration/19678-715609/.minikube/addons for local assets ...
I0920 19:21:55.263052 723137 filesync.go:126] Scanning /home/jenkins/minikube-integration/19678-715609/.minikube/files for local assets ...
I0920 19:21:55.263094 723137 start.go:296] duration metric: took 127.565352ms for postStartSetup
I0920 19:21:55.263474 723137 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-711398
I0920 19:21:55.281798 723137 profile.go:143] Saving config to /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/config.json ...
I0920 19:21:55.282096 723137 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0920 19:21:55.282159 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:55.299831 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:21:55.397218 723137 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0920 19:21:55.402069 723137 start.go:128] duration metric: took 10.623800204s to createHost
I0920 19:21:55.402093 723137 start.go:83] releasing machines lock for "addons-711398", held for 10.62394795s
I0920 19:21:55.402177 723137 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-711398
I0920 19:21:55.418979 723137 ssh_runner.go:195] Run: cat /version.json
I0920 19:21:55.418996 723137 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0920 19:21:55.419058 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:55.419099 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:21:55.440389 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:21:55.447911 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:21:55.540042 723137 ssh_runner.go:195] Run: systemctl --version
I0920 19:21:55.671136 723137 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0920 19:21:55.675552 723137 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0920 19:21:55.703261 723137 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0920 19:21:55.703353 723137 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0920 19:21:55.736315 723137 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0920 19:21:55.736383 723137 start.go:495] detecting cgroup driver to use...
I0920 19:21:55.736431 723137 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0920 19:21:55.736592 723137 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0920 19:21:55.752920 723137 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0920 19:21:55.762844 723137 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0920 19:21:55.773307 723137 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0920 19:21:55.773427 723137 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0920 19:21:55.783764 723137 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0920 19:21:55.794242 723137 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0920 19:21:55.804757 723137 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0920 19:21:55.814960 723137 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0920 19:21:55.824445 723137 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0920 19:21:55.834491 723137 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0920 19:21:55.844447 723137 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0920 19:21:55.854805 723137 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0920 19:21:55.863820 723137 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0920 19:21:55.872502 723137 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 19:21:55.957503 723137 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0920 19:21:56.064661 723137 start.go:495] detecting cgroup driver to use...
I0920 19:21:56.064706 723137 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0920 19:21:56.064769 723137 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0920 19:21:56.079337 723137 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0920 19:21:56.079426 723137 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0920 19:21:56.093925 723137 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0920 19:21:56.110324 723137 ssh_runner.go:195] Run: which cri-dockerd
I0920 19:21:56.114612 723137 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0920 19:21:56.124980 723137 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0920 19:21:56.144845 723137 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0920 19:21:56.246212 723137 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0920 19:21:56.341598 723137 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0920 19:21:56.341729 723137 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0920 19:21:56.363583 723137 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 19:21:56.462843 723137 ssh_runner.go:195] Run: sudo systemctl restart docker
I0920 19:21:56.740432 723137 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0920 19:21:56.753139 723137 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0920 19:21:56.765397 723137 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0920 19:21:56.862674 723137 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0920 19:21:56.957070 723137 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 19:21:57.048459 723137 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0920 19:21:57.063722 723137 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0920 19:21:57.075884 723137 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 19:21:57.158527 723137 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0920 19:21:57.226149 723137 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0920 19:21:57.226239 723137 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0920 19:21:57.229975 723137 start.go:563] Will wait 60s for crictl version
I0920 19:21:57.230038 723137 ssh_runner.go:195] Run: which crictl
I0920 19:21:57.233776 723137 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0920 19:21:57.270441 723137 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.2.1
RuntimeApiVersion: v1
I0920 19:21:57.270515 723137 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0920 19:21:57.293358 723137 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0920 19:21:57.320615 723137 out.go:235] * Preparing Kubernetes v1.31.1 on Docker 27.2.1 ...
I0920 19:21:57.320750 723137 cli_runner.go:164] Run: docker network inspect addons-711398 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0920 19:21:57.343824 723137 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0920 19:21:57.347657 723137 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0920 19:21:57.358868 723137 kubeadm.go:883] updating cluster {Name:addons-711398 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-711398 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0920 19:21:57.358994 723137 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0920 19:21:57.359058 723137 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0920 19:21:57.377404 723137 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0920 19:21:57.377425 723137 docker.go:615] Images already preloaded, skipping extraction
I0920 19:21:57.377494 723137 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0920 19:21:57.395122 723137 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0920 19:21:57.395143 723137 cache_images.go:84] Images are preloaded, skipping loading
I0920 19:21:57.395152 723137 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.1 docker true true} ...
I0920 19:21:57.395267 723137 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-711398 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.1 ClusterName:addons-711398 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0920 19:21:57.395331 723137 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0920 19:21:57.446501 723137 cni.go:84] Creating CNI manager for ""
I0920 19:21:57.446540 723137 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0920 19:21:57.446551 723137 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0920 19:21:57.446573 723137 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-711398 NodeName:addons-711398 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0920 19:21:57.446744 723137 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-711398"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0920 19:21:57.446825 723137 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.1
I0920 19:21:57.455776 723137 binaries.go:44] Found k8s binaries, skipping transfer
I0920 19:21:57.455848 723137 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0920 19:21:57.464817 723137 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0920 19:21:57.483136 723137 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0920 19:21:57.501157 723137 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0920 19:21:57.519635 723137 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0920 19:21:57.523365 723137 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0920 19:21:57.534249 723137 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 19:21:57.630642 723137 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0920 19:21:57.646531 723137 certs.go:68] Setting up /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398 for IP: 192.168.49.2
I0920 19:21:57.646596 723137 certs.go:194] generating shared ca certs ...
I0920 19:21:57.646630 723137 certs.go:226] acquiring lock for ca certs: {Name:mka146aeb8849fa662afd098460ee50b76cdcd3d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:57.646778 723137 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19678-715609/.minikube/ca.key
I0920 19:21:58.101679 723137 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19678-715609/.minikube/ca.crt ...
I0920 19:21:58.101711 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/ca.crt: {Name:mk69ee28a9294d68443622c5f147a8be9eb2e2c9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:58.101914 723137 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19678-715609/.minikube/ca.key ...
I0920 19:21:58.101930 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/ca.key: {Name:mk256dee300035c9b2b81542f780656aadb74bec Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:58.102570 723137 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19678-715609/.minikube/proxy-client-ca.key
I0920 19:21:58.361763 723137 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19678-715609/.minikube/proxy-client-ca.crt ...
I0920 19:21:58.361794 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/proxy-client-ca.crt: {Name:mk3cc193656d1fed6845639ce18a1d53df71f4fe Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:58.362556 723137 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19678-715609/.minikube/proxy-client-ca.key ...
I0920 19:21:58.362573 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/proxy-client-ca.key: {Name:mk7b22eaa279387750056e650cc130742f498b26 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:58.363135 723137 certs.go:256] generating profile certs ...
I0920 19:21:58.363204 723137 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/client.key
I0920 19:21:58.363224 723137 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/client.crt with IP's: []
I0920 19:21:58.544718 723137 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/client.crt ...
I0920 19:21:58.544750 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/client.crt: {Name:mk46d763dadcef5e2e8f9a813ad527b4958e2eae Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:58.544962 723137 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/client.key ...
I0920 19:21:58.544976 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/client.key: {Name:mk89fa1bda5da493f73eb566eed0799f6479246f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:58.545070 723137 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.key.3b3594dd
I0920 19:21:58.545096 723137 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.crt.3b3594dd with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0920 19:21:59.256455 723137 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.crt.3b3594dd ...
I0920 19:21:59.256487 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.crt.3b3594dd: {Name:mkf93a8399d411e005cb995b5473756fe79068d2 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:59.257249 723137 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.key.3b3594dd ...
I0920 19:21:59.257269 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.key.3b3594dd: {Name:mk34dae1678cd9d33f265d15cc56fec4b2987f6b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:59.257371 723137 certs.go:381] copying /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.crt.3b3594dd -> /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.crt
I0920 19:21:59.257458 723137 certs.go:385] copying /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.key.3b3594dd -> /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.key
I0920 19:21:59.257515 723137 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/proxy-client.key
I0920 19:21:59.257538 723137 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/proxy-client.crt with IP's: []
I0920 19:21:59.662051 723137 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/proxy-client.crt ...
I0920 19:21:59.662083 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/proxy-client.crt: {Name:mkf9ed1a1cbe47dfb4118289a791fe9d07070325 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:59.662793 723137 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/proxy-client.key ...
I0920 19:21:59.662813 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/proxy-client.key: {Name:mk9f93f8f45579477b9e1a14164c8861d4e2d6e1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:21:59.663529 723137 certs.go:484] found cert: /home/jenkins/minikube-integration/19678-715609/.minikube/certs/ca-key.pem (1679 bytes)
I0920 19:21:59.663576 723137 certs.go:484] found cert: /home/jenkins/minikube-integration/19678-715609/.minikube/certs/ca.pem (1078 bytes)
I0920 19:21:59.663601 723137 certs.go:484] found cert: /home/jenkins/minikube-integration/19678-715609/.minikube/certs/cert.pem (1123 bytes)
I0920 19:21:59.663628 723137 certs.go:484] found cert: /home/jenkins/minikube-integration/19678-715609/.minikube/certs/key.pem (1679 bytes)
I0920 19:21:59.664317 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0920 19:21:59.690160 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0920 19:21:59.715584 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0920 19:21:59.740820 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0920 19:21:59.766156 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0920 19:21:59.790742 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0920 19:21:59.816050 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0920 19:21:59.841551 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/profiles/addons-711398/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I0920 19:21:59.866259 723137 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19678-715609/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0920 19:21:59.891399 723137 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0920 19:21:59.910129 723137 ssh_runner.go:195] Run: openssl version
I0920 19:21:59.915900 723137 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0920 19:21:59.926060 723137 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0920 19:21:59.929543 723137 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 20 19:21 /usr/share/ca-certificates/minikubeCA.pem
I0920 19:21:59.929613 723137 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0920 19:21:59.936824 723137 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0920 19:21:59.946835 723137 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0920 19:21:59.950306 723137 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0920 19:21:59.950358 723137 kubeadm.go:392] StartCluster: {Name:addons-711398 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-711398 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0920 19:21:59.950545 723137 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0920 19:21:59.966637 723137 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0920 19:21:59.975741 723137 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0920 19:21:59.985067 723137 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0920 19:21:59.985132 723137 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0920 19:21:59.994501 723137 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0920 19:21:59.994522 723137 kubeadm.go:157] found existing configuration files:
I0920 19:21:59.994597 723137 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0920 19:22:00.005288 723137 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0920 19:22:00.005374 723137 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0920 19:22:00.064774 723137 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0920 19:22:00.096975 723137 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0920 19:22:00.097140 723137 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0920 19:22:00.135435 723137 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0920 19:22:00.169656 723137 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0920 19:22:00.169776 723137 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0920 19:22:00.184833 723137 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0920 19:22:00.203029 723137 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0920 19:22:00.203108 723137 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0920 19:22:00.220543 723137 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0920 19:22:00.309807 723137 kubeadm.go:310] [init] Using Kubernetes version: v1.31.1
I0920 19:22:00.309895 723137 kubeadm.go:310] [preflight] Running pre-flight checks
I0920 19:22:00.342830 723137 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0920 19:22:00.342946 723137 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1070-aws[0m
I0920 19:22:00.342992 723137 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0920 19:22:00.343097 723137 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0920 19:22:00.343181 723137 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0920 19:22:00.343253 723137 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0920 19:22:00.343333 723137 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0920 19:22:00.343384 723137 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0920 19:22:00.343490 723137 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0920 19:22:00.343576 723137 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0920 19:22:00.343654 723137 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0920 19:22:00.343732 723137 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0920 19:22:00.412748 723137 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0920 19:22:00.412867 723137 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0920 19:22:00.412969 723137 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0920 19:22:00.429535 723137 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0920 19:22:00.433801 723137 out.go:235] - Generating certificates and keys ...
I0920 19:22:00.433907 723137 kubeadm.go:310] [certs] Using existing ca certificate authority
I0920 19:22:00.434040 723137 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0920 19:22:01.151854 723137 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0920 19:22:02.085958 723137 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0920 19:22:02.587754 723137 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0920 19:22:02.715484 723137 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0920 19:22:02.981493 723137 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0920 19:22:02.981802 723137 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-711398 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0920 19:22:03.171622 723137 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0920 19:22:03.171963 723137 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-711398 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0920 19:22:03.392599 723137 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0920 19:22:03.845538 723137 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0920 19:22:04.307301 723137 kubeadm.go:310] [certs] Generating "sa" key and public key
I0920 19:22:04.307591 723137 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0920 19:22:04.631532 723137 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0920 19:22:05.347396 723137 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0920 19:22:06.535018 723137 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0920 19:22:06.853494 723137 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0920 19:22:07.642803 723137 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0920 19:22:07.643570 723137 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0920 19:22:07.646537 723137 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0920 19:22:07.649558 723137 out.go:235] - Booting up control plane ...
I0920 19:22:07.649656 723137 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0920 19:22:07.649731 723137 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0920 19:22:07.649809 723137 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0920 19:22:07.660474 723137 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0920 19:22:07.667481 723137 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0920 19:22:07.667536 723137 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0920 19:22:07.771828 723137 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0920 19:22:07.771980 723137 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0920 19:22:08.773239 723137 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.00167931s
I0920 19:22:08.773331 723137 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0920 19:22:15.776303 723137 kubeadm.go:310] [api-check] The API server is healthy after 7.002998254s
I0920 19:22:15.796684 723137 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0920 19:22:15.812586 723137 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0920 19:22:15.838528 723137 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0920 19:22:15.838726 723137 kubeadm.go:310] [mark-control-plane] Marking the node addons-711398 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0920 19:22:15.849813 723137 kubeadm.go:310] [bootstrap-token] Using token: gov4e8.ybo58bu4jqgk7v1j
I0920 19:22:15.852540 723137 out.go:235] - Configuring RBAC rules ...
I0920 19:22:15.852668 723137 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0920 19:22:15.857300 723137 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0920 19:22:15.865856 723137 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0920 19:22:15.869915 723137 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0920 19:22:15.875952 723137 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0920 19:22:15.880273 723137 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0920 19:22:16.184900 723137 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0920 19:22:16.611289 723137 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0920 19:22:17.183097 723137 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0920 19:22:17.184266 723137 kubeadm.go:310]
I0920 19:22:17.184342 723137 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0920 19:22:17.184351 723137 kubeadm.go:310]
I0920 19:22:17.184428 723137 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0920 19:22:17.184439 723137 kubeadm.go:310]
I0920 19:22:17.184464 723137 kubeadm.go:310] mkdir -p $HOME/.kube
I0920 19:22:17.184523 723137 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0920 19:22:17.184577 723137 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0920 19:22:17.184585 723137 kubeadm.go:310]
I0920 19:22:17.184638 723137 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0920 19:22:17.184651 723137 kubeadm.go:310]
I0920 19:22:17.184698 723137 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0920 19:22:17.184705 723137 kubeadm.go:310]
I0920 19:22:17.184757 723137 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0920 19:22:17.184838 723137 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0920 19:22:17.184910 723137 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0920 19:22:17.184918 723137 kubeadm.go:310]
I0920 19:22:17.185001 723137 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0920 19:22:17.185084 723137 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0920 19:22:17.185092 723137 kubeadm.go:310]
I0920 19:22:17.185175 723137 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token gov4e8.ybo58bu4jqgk7v1j \
I0920 19:22:17.185281 723137 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:3d9817bded5b2df7a22de97c06b991f8c482a20a289feb8315a94f93fe733f2a \
I0920 19:22:17.185313 723137 kubeadm.go:310] --control-plane
I0920 19:22:17.185323 723137 kubeadm.go:310]
I0920 19:22:17.185408 723137 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0920 19:22:17.185418 723137 kubeadm.go:310]
I0920 19:22:17.185499 723137 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token gov4e8.ybo58bu4jqgk7v1j \
I0920 19:22:17.185603 723137 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:3d9817bded5b2df7a22de97c06b991f8c482a20a289feb8315a94f93fe733f2a
I0920 19:22:17.190169 723137 kubeadm.go:310] W0920 19:22:00.288785 1807 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0920 19:22:17.190489 723137 kubeadm.go:310] W0920 19:22:00.290550 1807 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0920 19:22:17.190719 723137 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1070-aws\n", err: exit status 1
I0920 19:22:17.190833 723137 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0920 19:22:17.190862 723137 cni.go:84] Creating CNI manager for ""
I0920 19:22:17.190884 723137 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0920 19:22:17.195533 723137 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0920 19:22:17.198284 723137 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0920 19:22:17.207374 723137 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0920 19:22:17.226065 723137 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0920 19:22:17.226193 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 19:22:17.226276 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-711398 minikube.k8s.io/updated_at=2024_09_20T19_22_17_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=57d42ff8d541388826f5d9c37044129ec69c3d0a minikube.k8s.io/name=addons-711398 minikube.k8s.io/primary=true
I0920 19:22:17.367531 723137 ops.go:34] apiserver oom_adj: -16
I0920 19:22:17.367638 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 19:22:17.868108 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 19:22:18.368162 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 19:22:18.868572 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 19:22:19.367766 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 19:22:19.868638 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 19:22:20.367915 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 19:22:20.868495 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 19:22:21.368446 723137 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 19:22:21.468871 723137 kubeadm.go:1113] duration metric: took 4.242722067s to wait for elevateKubeSystemPrivileges
I0920 19:22:21.468896 723137 kubeadm.go:394] duration metric: took 21.518544137s to StartCluster
I0920 19:22:21.468912 723137 settings.go:142] acquiring lock: {Name:mk489fae9706e26496450fd05dacf08ba58ec1e6 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:22:21.469027 723137 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19678-715609/kubeconfig
I0920 19:22:21.469413 723137 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19678-715609/kubeconfig: {Name:mk865c6d1bfbd69f4aebff691c82d6c1986ead8f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 19:22:21.470242 723137 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0920 19:22:21.470266 723137 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0920 19:22:21.470503 723137 config.go:182] Loaded profile config "addons-711398": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0920 19:22:21.470552 723137 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0920 19:22:21.470643 723137 addons.go:69] Setting yakd=true in profile "addons-711398"
I0920 19:22:21.470663 723137 addons.go:234] Setting addon yakd=true in "addons-711398"
I0920 19:22:21.470690 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.471164 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.471648 723137 addons.go:69] Setting cloud-spanner=true in profile "addons-711398"
I0920 19:22:21.471668 723137 addons.go:234] Setting addon cloud-spanner=true in "addons-711398"
I0920 19:22:21.471692 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.471747 723137 addons.go:69] Setting metrics-server=true in profile "addons-711398"
I0920 19:22:21.471762 723137 addons.go:234] Setting addon metrics-server=true in "addons-711398"
I0920 19:22:21.471784 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.472143 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.472278 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.472606 723137 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-711398"
I0920 19:22:21.472625 723137 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-711398"
I0920 19:22:21.472647 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.473056 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.476184 723137 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-711398"
I0920 19:22:21.476250 723137 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-711398"
I0920 19:22:21.476280 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.477920 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.485029 723137 addons.go:69] Setting registry=true in profile "addons-711398"
I0920 19:22:21.485102 723137 addons.go:234] Setting addon registry=true in "addons-711398"
I0920 19:22:21.485152 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.485664 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.500150 723137 addons.go:69] Setting default-storageclass=true in profile "addons-711398"
I0920 19:22:21.500193 723137 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-711398"
I0920 19:22:21.500546 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.508202 723137 addons.go:69] Setting storage-provisioner=true in profile "addons-711398"
I0920 19:22:21.508287 723137 addons.go:234] Setting addon storage-provisioner=true in "addons-711398"
I0920 19:22:21.508358 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.508876 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.523958 723137 addons.go:69] Setting gcp-auth=true in profile "addons-711398"
I0920 19:22:21.524001 723137 mustload.go:65] Loading cluster: addons-711398
I0920 19:22:21.524267 723137 config.go:182] Loaded profile config "addons-711398": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0920 19:22:21.524546 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.534288 723137 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-711398"
I0920 19:22:21.534323 723137 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-711398"
I0920 19:22:21.534678 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.574831 723137 addons.go:69] Setting ingress=true in profile "addons-711398"
I0920 19:22:21.574872 723137 addons.go:234] Setting addon ingress=true in "addons-711398"
I0920 19:22:21.574919 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.575439 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.597985 723137 addons.go:69] Setting ingress-dns=true in profile "addons-711398"
I0920 19:22:21.598022 723137 addons.go:234] Setting addon ingress-dns=true in "addons-711398"
I0920 19:22:21.598075 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.598580 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.608147 723137 addons.go:69] Setting volcano=true in profile "addons-711398"
I0920 19:22:21.608185 723137 addons.go:234] Setting addon volcano=true in "addons-711398"
I0920 19:22:21.608223 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.608745 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.627817 723137 addons.go:69] Setting inspektor-gadget=true in profile "addons-711398"
I0920 19:22:21.627852 723137 addons.go:234] Setting addon inspektor-gadget=true in "addons-711398"
I0920 19:22:21.627892 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.628451 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.637098 723137 addons.go:69] Setting volumesnapshots=true in profile "addons-711398"
I0920 19:22:21.637142 723137 addons.go:234] Setting addon volumesnapshots=true in "addons-711398"
I0920 19:22:21.637180 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.637680 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.652467 723137 out.go:177] * Verifying Kubernetes components...
I0920 19:22:21.680728 723137 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0920 19:22:21.685847 723137 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0920 19:22:21.685880 723137 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0920 19:22:21.685956 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.700244 723137 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 19:22:21.708123 723137 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0920 19:22:21.711016 723137 addons.go:234] Setting addon default-storageclass=true in "addons-711398"
I0920 19:22:21.712994 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.711201 723137 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0920 19:22:21.711282 723137 out.go:177] - Using image docker.io/registry:2.8.3
I0920 19:22:21.711292 723137 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0920 19:22:21.712883 723137 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0920 19:22:21.712914 723137 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0920 19:22:21.713668 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.718803 723137 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-711398"
I0920 19:22:21.723910 723137 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0920 19:22:21.725650 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.728180 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.728742 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:21.745522 723137 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0920 19:22:21.745973 723137 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0920 19:22:21.748017 723137 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0920 19:22:21.728998 723137 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0920 19:22:21.762349 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0920 19:22:21.761914 723137 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0920 19:22:21.762550 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0920 19:22:21.762811 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.761926 723137 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0920 19:22:21.774899 723137 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0920 19:22:21.775796 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0920 19:22:21.776061 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.794468 723137 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0920 19:22:21.762749 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.797543 723137 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0920 19:22:21.797609 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0920 19:22:21.797715 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.835483 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:21.844170 723137 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0920 19:22:21.844320 723137 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0920 19:22:21.844359 723137 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0920 19:22:21.848448 723137 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0920 19:22:21.848604 723137 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0920 19:22:21.848651 723137 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0920 19:22:21.852844 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.877952 723137 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0920 19:22:21.883038 723137 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0920 19:22:21.890941 723137 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0920 19:22:21.901962 723137 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0920 19:22:21.901975 723137 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0920 19:22:21.904609 723137 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0920 19:22:21.906195 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0920 19:22:21.906264 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.911587 723137 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0920 19:22:21.911664 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0920 19:22:21.911771 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.931305 723137 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0920 19:22:21.934190 723137 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0920 19:22:21.934215 723137 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0920 19:22:21.934287 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.936325 723137 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0920 19:22:21.938977 723137 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0920 19:22:21.940163 723137 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0920 19:22:21.942685 723137 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0920 19:22:21.942708 723137 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0920 19:22:21.942789 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.952129 723137 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0920 19:22:21.952156 723137 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0920 19:22:21.952228 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.978877 723137 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0920 19:22:21.981745 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:21.984949 723137 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0920 19:22:21.984982 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0920 19:22:21.985051 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:21.991957 723137 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0920 19:22:21.995560 723137 out.go:177] - Using image docker.io/busybox:stable
I0920 19:22:22.000225 723137 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0920 19:22:22.000319 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0920 19:22:22.000431 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:22.028225 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.039714 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.040607 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.071257 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.079807 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.114319 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.120045 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.120736 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.124875 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.128415 723137 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0920 19:22:22.152303 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.155877 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
W0920 19:22:22.157144 723137 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0920 19:22:22.157177 723137 retry.go:31] will retry after 370.686783ms: ssh: handshake failed: EOF
W0920 19:22:22.157667 723137 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0920 19:22:22.157700 723137 retry.go:31] will retry after 188.036599ms: ssh: handshake failed: EOF
I0920 19:22:22.169132 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.169366 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:22.753184 723137 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0920 19:22:22.753210 723137 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0920 19:22:22.850117 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0920 19:22:22.995420 723137 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0920 19:22:22.995498 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0920 19:22:23.004718 723137 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0920 19:22:23.004800 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0920 19:22:23.010816 723137 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0920 19:22:23.010920 723137 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0920 19:22:23.014199 723137 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0920 19:22:23.014293 723137 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0920 19:22:23.031020 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0920 19:22:23.100793 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0920 19:22:23.175050 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0920 19:22:23.225822 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0920 19:22:23.237466 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0920 19:22:23.240388 723137 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0920 19:22:23.240474 723137 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0920 19:22:23.243677 723137 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0920 19:22:23.243758 723137 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0920 19:22:23.418104 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0920 19:22:23.534530 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0920 19:22:23.558242 723137 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0920 19:22:23.558323 723137 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0920 19:22:23.650134 723137 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0920 19:22:23.650215 723137 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0920 19:22:23.665189 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0920 19:22:23.679302 723137 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0920 19:22:23.679377 723137 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0920 19:22:23.687384 723137 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0920 19:22:23.687459 723137 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0920 19:22:23.810740 723137 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0920 19:22:23.810811 723137 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0920 19:22:23.838514 723137 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0920 19:22:23.838583 723137 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0920 19:22:23.879514 723137 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0920 19:22:23.879594 723137 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0920 19:22:23.890607 723137 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0920 19:22:23.890700 723137 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0920 19:22:23.972664 723137 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0920 19:22:23.972741 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0920 19:22:23.975074 723137 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (2.229074799s)
I0920 19:22:23.975148 723137 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0920 19:22:23.975256 723137 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.846822117s)
I0920 19:22:23.977184 723137 node_ready.go:35] waiting up to 6m0s for node "addons-711398" to be "Ready" ...
I0920 19:22:23.980846 723137 node_ready.go:49] node "addons-711398" has status "Ready":"True"
I0920 19:22:23.980876 723137 node_ready.go:38] duration metric: took 3.506242ms for node "addons-711398" to be "Ready" ...
I0920 19:22:23.980887 723137 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0920 19:22:23.991391 723137 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-sdxgq" in "kube-system" namespace to be "Ready" ...
I0920 19:22:24.037676 723137 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0920 19:22:24.037757 723137 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0920 19:22:24.049044 723137 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0920 19:22:24.049072 723137 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0920 19:22:24.195575 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0920 19:22:24.210992 723137 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0920 19:22:24.211016 723137 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0920 19:22:24.290237 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0920 19:22:24.365116 723137 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0920 19:22:24.365196 723137 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0920 19:22:24.391825 723137 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0920 19:22:24.391913 723137 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0920 19:22:24.453723 723137 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0920 19:22:24.453823 723137 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0920 19:22:24.480863 723137 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-711398" context rescaled to 1 replicas
I0920 19:22:24.703324 723137 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0920 19:22:24.703399 723137 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0920 19:22:24.741519 723137 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0920 19:22:24.741592 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0920 19:22:24.765867 723137 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0920 19:22:24.765947 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0920 19:22:24.921692 723137 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0920 19:22:24.921797 723137 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0920 19:22:25.020479 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0920 19:22:25.028617 723137 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0920 19:22:25.028710 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0920 19:22:25.062926 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0920 19:22:25.082442 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (2.232243717s)
I0920 19:22:25.132887 723137 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0920 19:22:25.132974 723137 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0920 19:22:25.413092 723137 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0920 19:22:25.413163 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0920 19:22:25.999441 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-sdxgq" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:26.027195 723137 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0920 19:22:26.027273 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0920 19:22:26.378115 723137 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0920 19:22:26.378199 723137 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0920 19:22:26.429569 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (3.398460588s)
I0920 19:22:27.046890 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0920 19:22:28.001524 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-sdxgq" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:28.901838 723137 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0920 19:22:28.901924 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:28.929516 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:29.209493 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (6.108663595s)
I0920 19:22:29.901687 723137 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0920 19:22:30.081717 723137 addons.go:234] Setting addon gcp-auth=true in "addons-711398"
I0920 19:22:30.081777 723137 host.go:66] Checking if "addons-711398" exists ...
I0920 19:22:30.082290 723137 cli_runner.go:164] Run: docker container inspect addons-711398 --format={{.State.Status}}
I0920 19:22:30.123649 723137 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0920 19:22:30.123722 723137 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-711398
I0920 19:22:30.154133 723137 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19678-715609/.minikube/machines/addons-711398/id_rsa Username:docker}
I0920 19:22:30.501505 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-sdxgq" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:32.502110 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-sdxgq" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:34.503774 723137 pod_ready.go:98] pod "coredns-7c65d6cfc9-sdxgq" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 19:22:34 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 19:22:22 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 19:22:22 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 19:22:22 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 19:22:22 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.49.2 HostIPs:[{IP:192.168.49.2
}] PodIP:10.244.0.2 PodIPs:[{IP:10.244.0.2}] StartTime:2024-09-20 19:22:22 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-09-20 19:22:23 +0000 UTC,FinishedAt:2024-09-20 19:22:33 +0000 UTC,ContainerID:docker://c2f4e989d93187e0930f326143bfcabea56025671d1547cccb537f118fd1b9de,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.3 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e ContainerID:docker://c2f4e989d93187e0930f326143bfcabea56025671d1547cccb537f118fd1b9de Started:0x400000fa30 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0x40017663c0} {Name:kube-api-access-wjqzh MountPath:/var/run/secrets/kubernetes.io/serviceaccount
ReadOnly:true RecursiveReadOnly:0x40017663d0}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
I0920 19:22:34.503996 723137 pod_ready.go:82] duration metric: took 10.512518616s for pod "coredns-7c65d6cfc9-sdxgq" in "kube-system" namespace to be "Ready" ...
E0920 19:22:34.504027 723137 pod_ready.go:67] WaitExtra: waitPodCondition: pod "coredns-7c65d6cfc9-sdxgq" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 19:22:34 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 19:22:22 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 19:22:22 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 19:22:22 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-20 19:22:22 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.4
9.2 HostIPs:[{IP:192.168.49.2}] PodIP:10.244.0.2 PodIPs:[{IP:10.244.0.2}] StartTime:2024-09-20 19:22:22 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-09-20 19:22:23 +0000 UTC,FinishedAt:2024-09-20 19:22:33 +0000 UTC,ContainerID:docker://c2f4e989d93187e0930f326143bfcabea56025671d1547cccb537f118fd1b9de,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.3 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e ContainerID:docker://c2f4e989d93187e0930f326143bfcabea56025671d1547cccb537f118fd1b9de Started:0x400000fa30 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0x40017663c0} {Name:kube-api-access-wjqzh MountPath:/var/run/secrets
/kubernetes.io/serviceaccount ReadOnly:true RecursiveReadOnly:0x40017663d0}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
I0920 19:22:34.504067 723137 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace to be "Ready" ...
I0920 19:22:34.905471 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.730337131s)
I0920 19:22:34.905590 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (11.679685902s)
I0920 19:22:34.905601 723137 addons.go:475] Verifying addon ingress=true in "addons-711398"
I0920 19:22:34.905883 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (11.668327621s)
I0920 19:22:34.905964 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (11.487784602s)
I0920 19:22:34.906000 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (11.371383803s)
I0920 19:22:34.906007 723137 addons.go:475] Verifying addon registry=true in "addons-711398"
I0920 19:22:34.906325 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (11.241063773s)
I0920 19:22:34.906555 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (10.710953896s)
I0920 19:22:34.906568 723137 addons.go:475] Verifying addon metrics-server=true in "addons-711398"
I0920 19:22:34.906658 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (10.616335068s)
I0920 19:22:34.906812 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (9.886229053s)
W0920 19:22:34.907123 723137 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0920 19:22:34.907142 723137 retry.go:31] will retry after 137.355606ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0920 19:22:34.906878 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (9.84387586s)
I0920 19:22:34.909585 723137 out.go:177] * Verifying ingress addon...
I0920 19:22:34.911339 723137 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-711398 service yakd-dashboard -n yakd-dashboard
I0920 19:22:34.911472 723137 out.go:177] * Verifying registry addon...
I0920 19:22:34.918896 723137 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0920 19:22:34.919902 723137 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0920 19:22:34.959109 723137 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0920 19:22:34.959194 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:34.960195 723137 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0920 19:22:34.960264 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:35.044920 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0920 19:22:35.434299 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:35.435403 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:35.944560 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:35.947813 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:35.975825 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (8.928849421s)
I0920 19:22:35.975910 723137 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-711398"
I0920 19:22:35.976278 723137 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.85260414s)
I0920 19:22:35.979103 723137 out.go:177] * Verifying csi-hostpath-driver addon...
I0920 19:22:35.979244 723137 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0920 19:22:35.983263 723137 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0920 19:22:35.983722 723137 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0920 19:22:35.985805 723137 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0920 19:22:35.985871 723137 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0920 19:22:35.992707 723137 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0920 19:22:35.992731 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:36.093964 723137 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0920 19:22:36.094052 723137 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0920 19:22:36.158566 723137 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0920 19:22:36.158633 723137 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0920 19:22:36.272998 723137 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0920 19:22:36.425286 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:36.425558 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:36.489245 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:36.510855 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:36.925764 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:36.926302 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:36.988359 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:37.396065 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.351040279s)
I0920 19:22:37.424831 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:37.426051 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:37.489044 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:37.807034 723137 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.533925293s)
I0920 19:22:37.810190 723137 addons.go:475] Verifying addon gcp-auth=true in "addons-711398"
I0920 19:22:37.813196 723137 out.go:177] * Verifying gcp-auth addon...
I0920 19:22:37.816718 723137 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0920 19:22:37.831723 723137 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0920 19:22:37.930875 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:37.931181 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:37.988393 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:38.425342 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:38.426228 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:38.511006 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:38.526304 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:38.924544 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:38.925622 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:38.989267 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:39.424223 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:39.425263 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:39.496407 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:39.924327 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:39.924865 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:39.988737 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:40.423590 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:40.424541 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:40.512132 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:40.524547 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:40.924713 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:40.926018 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:40.988474 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:41.424579 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:41.424898 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:41.501552 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:41.925393 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:41.926324 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:41.989622 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:42.424972 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:42.425938 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:42.488885 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:42.926548 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:42.927918 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:42.990656 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:43.011343 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:43.425582 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:43.426834 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:43.488610 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:43.923799 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:43.924783 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:43.988462 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:44.424338 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:44.425377 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:44.494302 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:44.925669 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:44.927287 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:44.989510 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:45.020512 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:45.427373 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:45.428377 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:45.488886 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:45.924307 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:45.925226 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:45.988804 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:46.423045 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:46.425865 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:46.490986 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:46.925033 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:46.927671 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:46.988784 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:47.424185 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:47.425633 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:47.489078 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:47.510048 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:47.926538 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:47.927498 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:47.988850 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:48.425836 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:48.426906 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:48.526560 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:48.924583 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:48.925099 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 19:22:48.989188 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:49.423487 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:49.424192 723137 kapi.go:107] duration metric: took 14.504291666s to wait for kubernetes.io/minikube-addons=registry ...
I0920 19:22:49.489231 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:49.511038 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:49.923446 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:49.988753 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:50.426152 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:50.488949 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:50.929151 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:50.988498 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:51.423899 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:51.488294 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:51.511093 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:51.923766 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:51.988227 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:52.423749 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:52.488776 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:52.923729 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:52.989054 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:53.423329 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:53.489889 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:53.923247 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:53.989095 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:54.011059 723137 pod_ready.go:103] pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace has status "Ready":"False"
I0920 19:22:54.423030 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:54.488705 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:54.923817 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:54.989821 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:55.423747 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:55.488032 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:55.924028 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:55.989228 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:56.011405 723137 pod_ready.go:93] pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace has status "Ready":"True"
I0920 19:22:56.011438 723137 pod_ready.go:82] duration metric: took 21.507331769s for pod "coredns-7c65d6cfc9-wkx75" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.011450 723137 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-711398" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.018812 723137 pod_ready.go:93] pod "etcd-addons-711398" in "kube-system" namespace has status "Ready":"True"
I0920 19:22:56.018854 723137 pod_ready.go:82] duration metric: took 7.395999ms for pod "etcd-addons-711398" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.018867 723137 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-711398" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.024829 723137 pod_ready.go:93] pod "kube-apiserver-addons-711398" in "kube-system" namespace has status "Ready":"True"
I0920 19:22:56.024867 723137 pod_ready.go:82] duration metric: took 5.991321ms for pod "kube-apiserver-addons-711398" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.024886 723137 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-711398" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.032379 723137 pod_ready.go:93] pod "kube-controller-manager-addons-711398" in "kube-system" namespace has status "Ready":"True"
I0920 19:22:56.032405 723137 pod_ready.go:82] duration metric: took 7.510622ms for pod "kube-controller-manager-addons-711398" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.032418 723137 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-mfhq6" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.039728 723137 pod_ready.go:93] pod "kube-proxy-mfhq6" in "kube-system" namespace has status "Ready":"True"
I0920 19:22:56.039756 723137 pod_ready.go:82] duration metric: took 7.32816ms for pod "kube-proxy-mfhq6" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.039768 723137 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-711398" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.408542 723137 pod_ready.go:93] pod "kube-scheduler-addons-711398" in "kube-system" namespace has status "Ready":"True"
I0920 19:22:56.408614 723137 pod_ready.go:82] duration metric: took 368.837436ms for pod "kube-scheduler-addons-711398" in "kube-system" namespace to be "Ready" ...
I0920 19:22:56.408641 723137 pod_ready.go:39] duration metric: took 32.427741025s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0920 19:22:56.408690 723137 api_server.go:52] waiting for apiserver process to appear ...
I0920 19:22:56.408778 723137 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0920 19:22:56.423446 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:56.437067 723137 api_server.go:72] duration metric: took 34.966767008s to wait for apiserver process to appear ...
I0920 19:22:56.437134 723137 api_server.go:88] waiting for apiserver healthz status ...
I0920 19:22:56.437170 723137 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0920 19:22:56.446641 723137 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0920 19:22:56.447710 723137 api_server.go:141] control plane version: v1.31.1
I0920 19:22:56.447764 723137 api_server.go:131] duration metric: took 10.609489ms to wait for apiserver health ...
I0920 19:22:56.447789 723137 system_pods.go:43] waiting for kube-system pods to appear ...
I0920 19:22:56.488836 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:56.616975 723137 system_pods.go:59] 17 kube-system pods found
I0920 19:22:56.617054 723137 system_pods.go:61] "coredns-7c65d6cfc9-wkx75" [146c07a3-8d88-4d28-939b-2957fd0149a8] Running
I0920 19:22:56.617083 723137 system_pods.go:61] "csi-hostpath-attacher-0" [8a937cd0-1930-43bc-9010-e3ba31744cc9] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0920 19:22:56.617127 723137 system_pods.go:61] "csi-hostpath-resizer-0" [1b574b1e-0912-4ac9-9d3d-ea851899a0a4] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0920 19:22:56.617160 723137 system_pods.go:61] "csi-hostpathplugin-h4mkl" [19a2a282-55e8-40b5-ae37-bf12079581fc] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0920 19:22:56.617184 723137 system_pods.go:61] "etcd-addons-711398" [47e53eda-7c09-462f-8b0e-96ac60cf38ec] Running
I0920 19:22:56.617208 723137 system_pods.go:61] "kube-apiserver-addons-711398" [1c65ff8c-35da-4071-8839-b21a66a7726b] Running
I0920 19:22:56.617244 723137 system_pods.go:61] "kube-controller-manager-addons-711398" [fffb5723-8da9-4e25-aa1c-2d27d4046b93] Running
I0920 19:22:56.617272 723137 system_pods.go:61] "kube-ingress-dns-minikube" [4b74ac90-44f6-410c-b476-f1c8a7d84b90] Running
I0920 19:22:56.617299 723137 system_pods.go:61] "kube-proxy-mfhq6" [dbfa9eee-c6dc-4c83-897e-7c31e823e7a8] Running
I0920 19:22:56.617317 723137 system_pods.go:61] "kube-scheduler-addons-711398" [c7a95707-c06b-434d-aa70-0bb07505c575] Running
I0920 19:22:56.617353 723137 system_pods.go:61] "metrics-server-84c5f94fbc-cwvt9" [2c61ac4c-97fc-4401-96cb-98c474378544] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0920 19:22:56.617381 723137 system_pods.go:61] "nvidia-device-plugin-daemonset-wqj2f" [706a55de-ce14-438b-bd2d-4793efdd30e7] Running
I0920 19:22:56.617405 723137 system_pods.go:61] "registry-66c9cd494c-84svt" [d2e45ba0-4b0a-4648-a233-1dfc5982c286] Running
I0920 19:22:56.617429 723137 system_pods.go:61] "registry-proxy-s7k45" [4fbca207-de93-4adb-baa8-2219f829573b] Running
I0920 19:22:56.617477 723137 system_pods.go:61] "snapshot-controller-56fcc65765-2xmsd" [10818109-c1c2-4475-b2aa-cdafb73ad5ae] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 19:22:56.617541 723137 system_pods.go:61] "snapshot-controller-56fcc65765-p8cms" [88bde4a5-50de-4db1-b9b6-f0dccace2981] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 19:22:56.617562 723137 system_pods.go:61] "storage-provisioner" [d1f5793e-a351-4154-9fda-390dc358bc7b] Running
I0920 19:22:56.617586 723137 system_pods.go:74] duration metric: took 169.776973ms to wait for pod list to return data ...
I0920 19:22:56.617619 723137 default_sa.go:34] waiting for default service account to be created ...
I0920 19:22:56.809313 723137 default_sa.go:45] found service account: "default"
I0920 19:22:56.809338 723137 default_sa.go:55] duration metric: took 191.695446ms for default service account to be created ...
I0920 19:22:56.809349 723137 system_pods.go:116] waiting for k8s-apps to be running ...
I0920 19:22:56.935432 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:56.989828 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:57.016394 723137 system_pods.go:86] 17 kube-system pods found
I0920 19:22:57.016470 723137 system_pods.go:89] "coredns-7c65d6cfc9-wkx75" [146c07a3-8d88-4d28-939b-2957fd0149a8] Running
I0920 19:22:57.016499 723137 system_pods.go:89] "csi-hostpath-attacher-0" [8a937cd0-1930-43bc-9010-e3ba31744cc9] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0920 19:22:57.016543 723137 system_pods.go:89] "csi-hostpath-resizer-0" [1b574b1e-0912-4ac9-9d3d-ea851899a0a4] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0920 19:22:57.016576 723137 system_pods.go:89] "csi-hostpathplugin-h4mkl" [19a2a282-55e8-40b5-ae37-bf12079581fc] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0920 19:22:57.016598 723137 system_pods.go:89] "etcd-addons-711398" [47e53eda-7c09-462f-8b0e-96ac60cf38ec] Running
I0920 19:22:57.016621 723137 system_pods.go:89] "kube-apiserver-addons-711398" [1c65ff8c-35da-4071-8839-b21a66a7726b] Running
I0920 19:22:57.016658 723137 system_pods.go:89] "kube-controller-manager-addons-711398" [fffb5723-8da9-4e25-aa1c-2d27d4046b93] Running
I0920 19:22:57.016689 723137 system_pods.go:89] "kube-ingress-dns-minikube" [4b74ac90-44f6-410c-b476-f1c8a7d84b90] Running
I0920 19:22:57.016711 723137 system_pods.go:89] "kube-proxy-mfhq6" [dbfa9eee-c6dc-4c83-897e-7c31e823e7a8] Running
I0920 19:22:57.016738 723137 system_pods.go:89] "kube-scheduler-addons-711398" [c7a95707-c06b-434d-aa70-0bb07505c575] Running
I0920 19:22:57.016776 723137 system_pods.go:89] "metrics-server-84c5f94fbc-cwvt9" [2c61ac4c-97fc-4401-96cb-98c474378544] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0920 19:22:57.016811 723137 system_pods.go:89] "nvidia-device-plugin-daemonset-wqj2f" [706a55de-ce14-438b-bd2d-4793efdd30e7] Running
I0920 19:22:57.016845 723137 system_pods.go:89] "registry-66c9cd494c-84svt" [d2e45ba0-4b0a-4648-a233-1dfc5982c286] Running
I0920 19:22:57.016894 723137 system_pods.go:89] "registry-proxy-s7k45" [4fbca207-de93-4adb-baa8-2219f829573b] Running
I0920 19:22:57.016924 723137 system_pods.go:89] "snapshot-controller-56fcc65765-2xmsd" [10818109-c1c2-4475-b2aa-cdafb73ad5ae] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 19:22:57.016951 723137 system_pods.go:89] "snapshot-controller-56fcc65765-p8cms" [88bde4a5-50de-4db1-b9b6-f0dccace2981] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 19:22:57.016974 723137 system_pods.go:89] "storage-provisioner" [d1f5793e-a351-4154-9fda-390dc358bc7b] Running
I0920 19:22:57.017009 723137 system_pods.go:126] duration metric: took 207.652299ms to wait for k8s-apps to be running ...
I0920 19:22:57.017037 723137 system_svc.go:44] waiting for kubelet service to be running ....
I0920 19:22:57.017133 723137 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0920 19:22:57.031921 723137 system_svc.go:56] duration metric: took 14.875385ms WaitForService to wait for kubelet
I0920 19:22:57.031949 723137 kubeadm.go:582] duration metric: took 35.561654655s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0920 19:22:57.031968 723137 node_conditions.go:102] verifying NodePressure condition ...
I0920 19:22:57.208986 723137 node_conditions.go:122] node storage ephemeral capacity is 203034800Ki
I0920 19:22:57.209025 723137 node_conditions.go:123] node cpu capacity is 2
I0920 19:22:57.209039 723137 node_conditions.go:105] duration metric: took 177.065174ms to run NodePressure ...
I0920 19:22:57.209051 723137 start.go:241] waiting for startup goroutines ...
I0920 19:22:57.423973 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:57.489235 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:57.924462 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:58.025797 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:58.424428 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:58.498694 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:58.926110 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:58.989870 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:59.432846 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:59.489679 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:22:59.923982 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:22:59.992103 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:00.426414 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:00.497531 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:00.924629 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:00.993184 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:01.423937 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:01.489293 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:01.924809 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:01.988491 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:02.423268 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:02.489664 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:02.923491 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:02.988707 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:03.423905 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:03.488454 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:03.923842 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:03.989419 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:04.424489 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:04.489609 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:04.924064 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:04.988713 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:05.425721 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:05.488613 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:05.925842 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:05.995290 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:06.423936 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:06.488463 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:06.924508 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:06.989318 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:07.426568 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:07.488845 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:07.923536 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:07.988937 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:08.423201 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:08.488532 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:08.927963 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:09.029843 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:09.423888 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:09.488670 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:09.924573 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:09.989768 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:10.426084 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:10.494864 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:10.923807 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:10.988247 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:11.423772 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:11.488917 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:11.922964 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:11.988407 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:12.426149 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:12.488730 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:12.923555 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:12.989240 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:13.425618 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:13.525680 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:13.924812 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:13.988374 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:14.428342 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:14.489121 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:14.924747 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:14.989218 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:15.425819 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:15.527663 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:15.923105 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:15.988370 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:16.424571 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:16.489007 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:16.923859 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:16.988923 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:17.425472 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:17.526206 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:17.924292 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:17.989083 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:18.423624 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:18.525237 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:18.922847 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:18.988307 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:19.423325 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:19.489146 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:19.924279 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:19.992550 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:20.423417 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:20.506202 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:20.923806 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:20.990245 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:21.424088 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:21.489314 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:21.925382 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:21.989047 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:22.423635 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:22.488028 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:22.923506 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:22.989827 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:23.423775 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:23.488389 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:23.938979 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:23.988572 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:24.424928 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:24.489199 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:24.923777 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:24.989232 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:25.424116 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:25.489181 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:25.932821 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:26.028783 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:26.423588 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:26.488962 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:26.923308 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:26.989430 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:27.428905 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:27.530264 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 19:23:27.923787 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:27.989065 723137 kapi.go:107] duration metric: took 52.005339361s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0920 19:23:28.423225 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:28.928167 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:29.423354 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:29.923371 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:30.423951 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:30.923275 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:31.423301 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:31.926340 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:32.423368 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:32.924520 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:33.423344 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:33.923735 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:34.423367 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:34.923449 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:35.423595 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:35.923282 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:36.422849 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:36.923739 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:37.428271 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:37.923695 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:38.424382 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:38.924416 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:39.424732 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:39.923627 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:40.423479 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:40.924430 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:41.423243 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:41.923985 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:42.424793 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:42.923786 723137 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 19:23:43.422789 723137 kapi.go:107] duration metric: took 1m8.503894524s to wait for app.kubernetes.io/name=ingress-nginx ...
I0920 19:23:59.820701 723137 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0920 19:23:59.820726 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:00.333843 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:00.820608 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:01.320643 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:01.821217 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:02.319773 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:02.820035 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:03.321076 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:03.821392 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:04.320240 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:04.820981 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:05.320303 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:05.820252 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:06.320543 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:06.819876 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:07.320750 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:07.820289 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:08.320111 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:08.820725 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:09.320847 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:09.820018 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:10.320993 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:10.820969 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:11.321337 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:11.819649 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:12.320156 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:12.820716 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:13.321044 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:13.820865 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:14.319994 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:14.820067 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:15.321052 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:15.822371 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:16.320390 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:16.820810 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:17.321132 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:17.819924 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:18.320874 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:18.821186 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:19.320292 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:19.820736 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:20.320719 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:20.820455 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:21.320260 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:21.820970 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:22.320374 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:22.819989 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:23.321164 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:23.820349 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:24.327977 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:24.820814 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:25.320832 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:25.820373 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:26.320027 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:26.822105 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:27.321417 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:27.820859 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:28.320291 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:28.819909 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:29.320804 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:29.820794 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:30.319932 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:30.820265 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:31.320189 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:31.820260 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:32.320665 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:32.820150 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:33.320395 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:33.820604 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:34.320833 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:34.820254 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:35.320939 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:35.820311 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:36.319875 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:36.825879 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:37.320519 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:37.820673 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:38.320351 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:38.819819 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:39.321173 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:39.820300 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:40.319998 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:40.821966 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:41.321272 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:41.824393 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:42.320216 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:42.819715 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:43.320404 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:43.819974 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:44.321261 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:44.821502 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:45.320668 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:45.820759 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:46.321279 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:46.819834 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:47.320897 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:47.819972 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:48.321078 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:48.820703 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:49.320529 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:49.821274 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:50.320195 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:50.820005 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:51.320207 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:51.820673 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:52.319919 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:52.820029 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:53.321138 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:53.820798 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:54.320476 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:54.820763 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:55.320674 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:55.820893 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:56.320151 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:56.820865 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:57.320777 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:57.820966 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:58.321309 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:58.820025 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:59.321111 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:24:59.821489 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:00.321238 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:00.820899 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:01.320764 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:01.821085 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:02.321219 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:02.820569 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:03.320395 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:03.819726 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:04.320156 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:04.820248 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:05.320112 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:05.821192 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:06.320792 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:06.819959 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:07.320753 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:07.821170 723137 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 19:25:08.320686 723137 kapi.go:107] duration metric: took 2m30.503965778s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0920 19:25:08.323479 723137 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-711398 cluster.
I0920 19:25:08.326671 723137 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0920 19:25:08.329370 723137 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0920 19:25:08.332210 723137 out.go:177] * Enabled addons: nvidia-device-plugin, storage-provisioner, storage-provisioner-rancher, volcano, cloud-spanner, ingress-dns, metrics-server, inspektor-gadget, yakd, default-storageclass, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0920 19:25:08.334899 723137 addons.go:510] duration metric: took 2m46.864347589s for enable addons: enabled=[nvidia-device-plugin storage-provisioner storage-provisioner-rancher volcano cloud-spanner ingress-dns metrics-server inspektor-gadget yakd default-storageclass volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0920 19:25:08.334950 723137 start.go:246] waiting for cluster config update ...
I0920 19:25:08.334972 723137 start.go:255] writing updated cluster config ...
I0920 19:25:08.335280 723137 ssh_runner.go:195] Run: rm -f paused
I0920 19:25:08.755213 723137 start.go:600] kubectl: 1.31.1, cluster: 1.31.1 (minor skew: 0)
I0920 19:25:08.758521 723137 out.go:177] * Done! kubectl is now configured to use "addons-711398" cluster and "default" namespace by default
==> Docker <==
Sep 20 19:34:34 addons-711398 dockerd[1280]: time="2024-09-20T19:34:34.797297476Z" level=info msg="ignoring event" container=2d29101c51549ecd00266081d2bd4a2aab2d0b2ee5ef6feaea94114d47d9ed5e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:34:34 addons-711398 dockerd[1280]: time="2024-09-20T19:34:34.817313571Z" level=info msg="ignoring event" container=7ee5628d32d2ab468d6c8a0c0ff0a56f3d727b0c36cefe0b405d4cd092e3b093 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:34:35 addons-711398 cri-dockerd[1536]: time="2024-09-20T19:34:35Z" level=info msg="Stop pulling image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0@sha256:03e677e1cf9d2c9bea454e3dbcbcef20b3022e987534a2874eb1abc5bc3e73ec: Status: Image is up to date for ghcr.io/inspektor-gadget/inspektor-gadget@sha256:03e677e1cf9d2c9bea454e3dbcbcef20b3022e987534a2874eb1abc5bc3e73ec"
Sep 20 19:34:37 addons-711398 dockerd[1280]: time="2024-09-20T19:34:37.092789084Z" level=info msg="ignoring event" container=d236dd6bb94bbbf59152d56092d24df67ba04978f85ad6a6517400e51dd1d6ad module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:34:41 addons-711398 dockerd[1280]: time="2024-09-20T19:34:41.314142297Z" level=info msg="ignoring event" container=0ecb5a23c8f61658c2f69090da93fb8148ab0fa15ab16607b3be2614f5e11a8b module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:34:41 addons-711398 dockerd[1280]: time="2024-09-20T19:34:41.491624115Z" level=info msg="ignoring event" container=53711243fb3a331b602f8f22aa3078231d02c28848bac1ca345718a67c77974e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:34:42 addons-711398 cri-dockerd[1536]: time="2024-09-20T19:34:42Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/e74930821d151aa178ef0ab1e682ca456ed20698a08b10b2e663c9fa924cfe3a/resolv.conf as [nameserver 10.96.0.10 search local-path-storage.svc.cluster.local svc.cluster.local cluster.local us-east-2.compute.internal options ndots:5]"
Sep 20 19:34:42 addons-711398 dockerd[1280]: time="2024-09-20T19:34:42.420791053Z" level=warning msg="reference for unknown type: " digest="sha256:3fbc632167424a6d997e74f52b878d7cc478225cffac6bc977eedfe51c7f4e79" remote="docker.io/library/busybox@sha256:3fbc632167424a6d997e74f52b878d7cc478225cffac6bc977eedfe51c7f4e79"
Sep 20 19:34:42 addons-711398 cri-dockerd[1536]: time="2024-09-20T19:34:42Z" level=info msg="Stop pulling image docker.io/busybox:stable@sha256:3fbc632167424a6d997e74f52b878d7cc478225cffac6bc977eedfe51c7f4e79: Status: Downloaded newer image for busybox@sha256:3fbc632167424a6d997e74f52b878d7cc478225cffac6bc977eedfe51c7f4e79"
Sep 20 19:34:43 addons-711398 dockerd[1280]: time="2024-09-20T19:34:43.089840272Z" level=info msg="ignoring event" container=c7762e73321c298423a81db357637945f8c3a226386b9af9f7101d6e5be8ae7c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:34:44 addons-711398 dockerd[1280]: time="2024-09-20T19:34:44.413061213Z" level=info msg="ignoring event" container=e74930821d151aa178ef0ab1e682ca456ed20698a08b10b2e663c9fa924cfe3a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:34:46 addons-711398 cri-dockerd[1536]: time="2024-09-20T19:34:46Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/996a1d75cb94210a2acca8efa688515cf74b6123bea3d4144d29757f52cdba7a/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local us-east-2.compute.internal options ndots:5]"
Sep 20 19:34:47 addons-711398 cri-dockerd[1536]: time="2024-09-20T19:34:47Z" level=info msg="Stop pulling image busybox:stable: Status: Downloaded newer image for busybox:stable"
Sep 20 19:34:47 addons-711398 dockerd[1280]: time="2024-09-20T19:34:47.345593737Z" level=info msg="ignoring event" container=8f6e27093614c60a9883dc5860fc67adb2b8aa2b7402a74eee86fe1f486f1792 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:34:49 addons-711398 dockerd[1280]: time="2024-09-20T19:34:49.542942939Z" level=info msg="ignoring event" container=996a1d75cb94210a2acca8efa688515cf74b6123bea3d4144d29757f52cdba7a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:34:49 addons-711398 dockerd[1280]: time="2024-09-20T19:34:49.721290138Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 20 19:34:49 addons-711398 dockerd[1280]: time="2024-09-20T19:34:49.723952973Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 20 19:34:50 addons-711398 cri-dockerd[1536]: time="2024-09-20T19:34:50Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/320fbcc25d086789d2f69cff8245784997a830201ad2d228faf2ddc1eabc24dd/resolv.conf as [nameserver 10.96.0.10 search local-path-storage.svc.cluster.local svc.cluster.local cluster.local us-east-2.compute.internal options ndots:5]"
Sep 20 19:34:51 addons-711398 dockerd[1280]: time="2024-09-20T19:34:51.167419695Z" level=info msg="ignoring event" container=7feea648c1fd1eafa1e03da6fb34b88c4518267d0c6f291c2d6bee7780aaf4ff module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:34:52 addons-711398 dockerd[1280]: time="2024-09-20T19:34:52.591349849Z" level=info msg="ignoring event" container=320fbcc25d086789d2f69cff8245784997a830201ad2d228faf2ddc1eabc24dd module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:35:05 addons-711398 dockerd[1280]: time="2024-09-20T19:35:05.168868614Z" level=info msg="ignoring event" container=5ce24c30324644d9885da4a436fddb8de565634d76ff26e54ed48c701348a4d5 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:35:05 addons-711398 dockerd[1280]: time="2024-09-20T19:35:05.817673849Z" level=info msg="ignoring event" container=f8330d199113ddd1713739370a5f6157c8a097384c87635281582db2e6df1481 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:35:05 addons-711398 dockerd[1280]: time="2024-09-20T19:35:05.937452841Z" level=info msg="ignoring event" container=c8ddd2e4784afd8f3ed04ab0294c18ddc8d819776f9b2f7cc18c6fb4fd526e2c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:35:06 addons-711398 dockerd[1280]: time="2024-09-20T19:35:06.059903292Z" level=info msg="ignoring event" container=ddeeebb4aa51f0ef66301296cee0c5ad999ec876d63ed1c728ad744604d508ba module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 19:35:06 addons-711398 dockerd[1280]: time="2024-09-20T19:35:06.197559740Z" level=info msg="ignoring event" container=4bbe09082ceb686bb0d488dbd36a4f09f2cce8cc55fc07a83dd0a1b4d56d40e9 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
7feea648c1fd1 fc9db2894f4e4 17 seconds ago Exited helper-pod 0 320fbcc25d086 helper-pod-delete-pvc-9a9bf7c2-70be-4ebd-8920-1988957db53e
8f6e27093614c busybox@sha256:c230832bd3b0be59a6c47ed64294f9ce71e91b327957920b6929a0caa8353140 20 seconds ago Exited busybox 0 996a1d75cb942 test-local-path
c7762e73321c2 busybox@sha256:3fbc632167424a6d997e74f52b878d7cc478225cffac6bc977eedfe51c7f4e79 25 seconds ago Exited helper-pod 0 e74930821d151 helper-pod-create-pvc-9a9bf7c2-70be-4ebd-8920-1988957db53e
d236dd6bb94bb ghcr.io/inspektor-gadget/inspektor-gadget@sha256:03e677e1cf9d2c9bea454e3dbcbcef20b3022e987534a2874eb1abc5bc3e73ec 32 seconds ago Exited gadget 7 45c2e44bacafd gadget-hp9tl
1443146b883ce gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 10 minutes ago Running gcp-auth 0 284d18d0bc73e gcp-auth-89d5ffd79-pjkck
a8ebfa27beee5 registry.k8s.io/ingress-nginx/controller@sha256:d5f8217feeac4887cb1ed21f27c2674e58be06bd8f5184cacea2a69abaf78dce 11 minutes ago Running controller 0 7399d2b7f25e4 ingress-nginx-controller-bc57996ff-bttnp
8cbf5f6e7b549 420193b27261a 11 minutes ago Exited patch 1 1a5c846bd0818 ingress-nginx-admission-patch-n2285
6c3796126af42 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 c1e2c04149434 ingress-nginx-admission-create-7mls9
4790c00475ce1 rancher/local-path-provisioner@sha256:e34c88ae0affb1cdefbb874140d6339d4a27ec4ee420ae8199cd839997b05246 12 minutes ago Running local-path-provisioner 0 479fc5ca9ddfa local-path-provisioner-86d989889c-kd5bv
cff85728082c6 registry.k8s.io/metrics-server/metrics-server@sha256:ffcb2bf004d6aa0a17d90e0247cf94f2865c8901dcab4427034c341951c239f9 12 minutes ago Running metrics-server 0 915dca5d2307e metrics-server-84c5f94fbc-cwvt9
6122dfaa35254 gcr.io/k8s-minikube/minikube-ingress-dns@sha256:4211a1de532376c881851542238121b26792225faa36a7b02dccad88fd05797c 12 minutes ago Running minikube-ingress-dns 0 bc6978c099a95 kube-ingress-dns-minikube
85e0a9205042f gcr.io/cloud-spanner-emulator/emulator@sha256:636fdfc528824bae5f0ea2eca6ae307fe81092f05ec21038008bc0d6100e52fc 12 minutes ago Running cloud-spanner-emulator 0 76f7ff38b323f cloud-spanner-emulator-769b77f747-g4968
dcbbde731377a ba04bb24b9575 12 minutes ago Running storage-provisioner 0 ec9228fda243b storage-provisioner
62308f2fbb336 2f6c962e7b831 12 minutes ago Running coredns 0 8b736fcdb1c85 coredns-7c65d6cfc9-wkx75
060c8c64d4225 24a140c548c07 12 minutes ago Running kube-proxy 0 89c24f17713f3 kube-proxy-mfhq6
3a5df68bb4480 27e3830e14027 12 minutes ago Running etcd 0 4055336d587e2 etcd-addons-711398
23d47197484db 7f8aa378bb47d 12 minutes ago Running kube-scheduler 0 a2a2b7d9fca75 kube-scheduler-addons-711398
cc9b700d2590d 279f381cb3736 12 minutes ago Running kube-controller-manager 0 6ffa0b97ce81a kube-controller-manager-addons-711398
c13c74eade154 d3f53a98c0a9d 12 minutes ago Running kube-apiserver 0 1db898b9f0ca2 kube-apiserver-addons-711398
==> controller_ingress [a8ebfa27beee] <==
W0920 19:23:43.064743 7 client_config.go:659] Neither --kubeconfig nor --master was specified. Using the inClusterConfig. This might not work.
I0920 19:23:43.065064 7 main.go:205] "Creating API client" host="https://10.96.0.1:443"
I0920 19:23:43.077472 7 main.go:248] "Running in Kubernetes cluster" major="1" minor="31" git="v1.31.1" state="clean" commit="948afe5ca072329a73c8e79ed5938717a5cb3d21" platform="linux/arm64"
I0920 19:23:44.061825 7 main.go:101] "SSL fake certificate created" file="/etc/ingress-controller/ssl/default-fake-certificate.pem"
I0920 19:23:44.082127 7 ssl.go:535] "loading tls certificate" path="/usr/local/certificates/cert" key="/usr/local/certificates/key"
I0920 19:23:44.094506 7 nginx.go:271] "Starting NGINX Ingress controller"
I0920 19:23:44.113454 7 event.go:377] Event(v1.ObjectReference{Kind:"ConfigMap", Namespace:"ingress-nginx", Name:"ingress-nginx-controller", UID:"ba8ec272-bb20-4175-beff-a07133e0c4aa", APIVersion:"v1", ResourceVersion:"698", FieldPath:""}): type: 'Normal' reason: 'CREATE' ConfigMap ingress-nginx/ingress-nginx-controller
I0920 19:23:44.131591 7 event.go:377] Event(v1.ObjectReference{Kind:"ConfigMap", Namespace:"ingress-nginx", Name:"tcp-services", UID:"27e10ab0-8982-43c2-986f-b93bc500847d", APIVersion:"v1", ResourceVersion:"699", FieldPath:""}): type: 'Normal' reason: 'CREATE' ConfigMap ingress-nginx/tcp-services
I0920 19:23:44.141224 7 event.go:377] Event(v1.ObjectReference{Kind:"ConfigMap", Namespace:"ingress-nginx", Name:"udp-services", UID:"38ec8c9e-3006-48a5-9910-103dcd0e7a5c", APIVersion:"v1", ResourceVersion:"700", FieldPath:""}): type: 'Normal' reason: 'CREATE' ConfigMap ingress-nginx/udp-services
I0920 19:23:45.296232 7 nginx.go:317] "Starting NGINX process"
I0920 19:23:45.296434 7 leaderelection.go:250] attempting to acquire leader lease ingress-nginx/ingress-nginx-leader...
I0920 19:23:45.299135 7 nginx.go:337] "Starting validation webhook" address=":8443" certPath="/usr/local/certificates/cert" keyPath="/usr/local/certificates/key"
I0920 19:23:45.299335 7 controller.go:193] "Configuration changes detected, backend reload required"
I0920 19:23:45.342721 7 leaderelection.go:260] successfully acquired lease ingress-nginx/ingress-nginx-leader
I0920 19:23:45.343113 7 status.go:85] "New leader elected" identity="ingress-nginx-controller-bc57996ff-bttnp"
I0920 19:23:45.352708 7 controller.go:213] "Backend successfully reloaded"
I0920 19:23:45.352905 7 controller.go:224] "Initial sync, sleeping for 1 second"
I0920 19:23:45.353062 7 event.go:377] Event(v1.ObjectReference{Kind:"Pod", Namespace:"ingress-nginx", Name:"ingress-nginx-controller-bc57996ff-bttnp", UID:"62c6c992-731f-4ffa-b530-a98b5b5103e2", APIVersion:"v1", ResourceVersion:"1281", FieldPath:""}): type: 'Normal' reason: 'RELOAD' NGINX reload triggered due to a change in configuration
I0920 19:23:45.431169 7 status.go:219] "POD is not ready" pod="ingress-nginx/ingress-nginx-controller-bc57996ff-bttnp" node="addons-711398"
Build: 46e76e5916813cfca2a9b0bfdc34b69a0000f6b9
Repository: https://github.com/kubernetes/ingress-nginx
nginx version: nginx/1.25.5
-------------------------------------------------------------------------------
==> coredns [62308f2fbb33] <==
Trace[274955516]: [30.001618174s] [30.001618174s] END
[INFO] plugin/kubernetes: Trace[986262570]: "Reflector ListAndWatch" name:pkg/mod/k8s.io/client-go@v0.29.3/tools/cache/reflector.go:229 (20-Sep-2024 19:22:23.618) (total time: 30000ms):
Trace[986262570]: ---"Objects listed" error:Get "https://10.96.0.1:443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.96.0.1:443: i/o timeout 30000ms (19:22:53.619)
Trace[986262570]: [30.00086798s] [30.00086798s] END
[ERROR] plugin/kubernetes: pkg/mod/k8s.io/client-go@v0.29.3/tools/cache/reflector.go:229: Failed to watch *v1.Namespace: failed to list *v1.Namespace: Get "https://10.96.0.1:443/api/v1/namespaces?limit=500&resourceVersion=0": dial tcp 10.96.0.1:443: i/o timeout
[ERROR] plugin/kubernetes: pkg/mod/k8s.io/client-go@v0.29.3/tools/cache/reflector.go:229: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.96.0.1:443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.96.0.1:443: i/o timeout
[INFO] plugin/kubernetes: pkg/mod/k8s.io/client-go@v0.29.3/tools/cache/reflector.go:229: failed to list *v1.EndpointSlice: Get "https://10.96.0.1:443/apis/discovery.k8s.io/v1/endpointslices?limit=500&resourceVersion=0": dial tcp 10.96.0.1:443: i/o timeout
[INFO] plugin/kubernetes: Trace[2068232413]: "Reflector ListAndWatch" name:pkg/mod/k8s.io/client-go@v0.29.3/tools/cache/reflector.go:229 (20-Sep-2024 19:22:23.618) (total time: 30000ms):
Trace[2068232413]: ---"Objects listed" error:Get "https://10.96.0.1:443/apis/discovery.k8s.io/v1/endpointslices?limit=500&resourceVersion=0": dial tcp 10.96.0.1:443: i/o timeout 30000ms (19:22:53.619)
Trace[2068232413]: [30.000736348s] [30.000736348s] END
[ERROR] plugin/kubernetes: pkg/mod/k8s.io/client-go@v0.29.3/tools/cache/reflector.go:229: Failed to watch *v1.EndpointSlice: failed to list *v1.EndpointSlice: Get "https://10.96.0.1:443/apis/discovery.k8s.io/v1/endpointslices?limit=500&resourceVersion=0": dial tcp 10.96.0.1:443: i/o timeout
[INFO] Reloading
[INFO] plugin/reload: Running configuration SHA512 = 05e3eaddc414b2d71a69b2e2bc6f2681fc1f4d04bcdd3acc1a41457bb7db518208b95ddfc4c9fffedc59c25a8faf458be1af4915a4a3c0d6777cb7a346bc5d86
[INFO] Reloading complete
[INFO] 127.0.0.1:41072 - 576 "HINFO IN 7204552164135868152.7786840848119861607. udp 57 false 512" NXDOMAIN qr,rd,ra 57 0.037833317s
[INFO] 10.244.0.25:55282 - 35073 "A IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000386198s
[INFO] 10.244.0.25:35547 - 26219 "AAAA IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000541944s
[INFO] 10.244.0.25:51421 - 30851 "AAAA IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000105819s
[INFO] 10.244.0.25:56747 - 11012 "A IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000288535s
[INFO] 10.244.0.25:53372 - 3255 "AAAA IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000155393s
[INFO] 10.244.0.25:42882 - 8912 "A IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000087252s
[INFO] 10.244.0.25:58952 - 23561 "AAAA IN storage.googleapis.com.us-east-2.compute.internal. udp 78 false 1232" NXDOMAIN qr,rd,ra 67 0.00258796s
[INFO] 10.244.0.25:54367 - 55125 "A IN storage.googleapis.com.us-east-2.compute.internal. udp 78 false 1232" NXDOMAIN qr,rd,ra 67 0.002294109s
[INFO] 10.244.0.25:54268 - 57761 "AAAA IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 240 0.002267475s
[INFO] 10.244.0.25:51107 - 5993 "A IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 610 0.001940804s
==> describe nodes <==
Name: addons-711398
Roles: control-plane
Labels: beta.kubernetes.io/arch=arm64
beta.kubernetes.io/os=linux
kubernetes.io/arch=arm64
kubernetes.io/hostname=addons-711398
kubernetes.io/os=linux
minikube.k8s.io/commit=57d42ff8d541388826f5d9c37044129ec69c3d0a
minikube.k8s.io/name=addons-711398
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_20T19_22_17_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-711398
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Fri, 20 Sep 2024 19:22:13 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-711398
AcquireTime: <unset>
RenewTime: Fri, 20 Sep 2024 19:35:03 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Fri, 20 Sep 2024 19:34:52 +0000 Fri, 20 Sep 2024 19:22:10 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Fri, 20 Sep 2024 19:34:52 +0000 Fri, 20 Sep 2024 19:22:10 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Fri, 20 Sep 2024 19:34:52 +0000 Fri, 20 Sep 2024 19:22:10 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Fri, 20 Sep 2024 19:34:52 +0000 Fri, 20 Sep 2024 19:22:13 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-711398
Capacity:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022304Ki
pods: 110
Allocatable:
cpu: 2
ephemeral-storage: 203034800Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
hugepages-32Mi: 0
hugepages-64Ki: 0
memory: 8022304Ki
pods: 110
System Info:
Machine ID: 21e1fd8952384521b2369f9b1931dd39
System UUID: f977aaa1-dba3-4af0-92ab-c521c8270934
Boot ID: 32c222cc-d06c-4f68-9fc3-59cd35d0dbd2
Kernel Version: 5.15.0-1070-aws
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: arm64
Container Runtime Version: docker://27.2.1
Kubelet Version: v1.31.1
Kube-Proxy Version: v1.31.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (15 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m17s
default cloud-spanner-emulator-769b77f747-g4968 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
gadget gadget-hp9tl 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
gcp-auth gcp-auth-89d5ffd79-pjkck 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
ingress-nginx ingress-nginx-controller-bc57996ff-bttnp 100m (5%) 0 (0%) 90Mi (1%) 0 (0%) 12m
kube-system coredns-7c65d6cfc9-wkx75 100m (5%) 0 (0%) 70Mi (0%) 170Mi (2%) 12m
kube-system etcd-addons-711398 100m (5%) 0 (0%) 100Mi (1%) 0 (0%) 12m
kube-system kube-apiserver-addons-711398 250m (12%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-711398 200m (10%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-ingress-dns-minikube 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-mfhq6 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-711398 100m (5%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system metrics-server-84c5f94fbc-cwvt9 100m (5%) 0 (0%) 200Mi (2%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
local-path-storage local-path-provisioner-86d989889c-kd5bv 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 950m (47%) 0 (0%)
memory 460Mi (5%) 170Mi (2%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
hugepages-32Mi 0 (0%) 0 (0%)
hugepages-64Ki 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal NodeHasSufficientMemory 12m (x8 over 12m) kubelet Node addons-711398 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m (x7 over 12m) kubelet Node addons-711398 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m (x7 over 12m) kubelet Node addons-711398 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-711398 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-711398 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-711398 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-711398 event: Registered Node addons-711398 in Controller
==> dmesg <==
[Sep20 18:55] overlayfs: '/var/lib/containers/storage/overlay/l/Q2QJNMTVZL6GMULS36RA5ZJGSA' not a directory
==> etcd [3a5df68bb448] <==
{"level":"info","ts":"2024-09-20T19:22:09.964874Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc switched to configuration voters=(12593026477526642892)"}
{"level":"info","ts":"2024-09-20T19:22:09.965218Z","caller":"membership/cluster.go:421","msg":"added member","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","added-peer-id":"aec36adc501070cc","added-peer-peer-urls":["https://192.168.49.2:2380"]}
{"level":"info","ts":"2024-09-20T19:22:10.684122Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc is starting a new election at term 1"}
{"level":"info","ts":"2024-09-20T19:22:10.684366Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became pre-candidate at term 1"}
{"level":"info","ts":"2024-09-20T19:22:10.684507Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgPreVoteResp from aec36adc501070cc at term 1"}
{"level":"info","ts":"2024-09-20T19:22:10.684642Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became candidate at term 2"}
{"level":"info","ts":"2024-09-20T19:22:10.684748Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-20T19:22:10.684878Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-09-20T19:22:10.684963Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-20T19:22:10.688275Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-711398 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-09-20T19:22:10.688568Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-20T19:22:10.689047Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T19:22:10.692116Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-20T19:22:10.693345Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-20T19:22:10.705338Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-09-20T19:22:10.700502Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T19:22:10.701670Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-20T19:22:10.702236Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-20T19:22:10.712660Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T19:22:10.713901Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-20T19:22:10.723183Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-20T19:22:10.728131Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T19:32:11.098166Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1887}
{"level":"info","ts":"2024-09-20T19:32:11.196142Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1887,"took":"97.112102ms","hash":126441347,"current-db-size-bytes":8859648,"current-db-size":"8.9 MB","current-db-size-in-use-bytes":4943872,"current-db-size-in-use":"4.9 MB"}
{"level":"info","ts":"2024-09-20T19:32:11.196203Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":126441347,"revision":1887,"compact-revision":-1}
==> gcp-auth [1443146b883c] <==
2024/09/20 19:25:07 GCP Auth Webhook started!
2024/09/20 19:25:25 Ready to marshal response ...
2024/09/20 19:25:25 Ready to write response ...
2024/09/20 19:25:25 Ready to marshal response ...
2024/09/20 19:25:25 Ready to write response ...
2024/09/20 19:25:50 Ready to marshal response ...
2024/09/20 19:25:50 Ready to write response ...
2024/09/20 19:25:50 Ready to marshal response ...
2024/09/20 19:25:50 Ready to write response ...
2024/09/20 19:25:50 Ready to marshal response ...
2024/09/20 19:25:50 Ready to write response ...
2024/09/20 19:34:04 Ready to marshal response ...
2024/09/20 19:34:04 Ready to write response ...
2024/09/20 19:34:05 Ready to marshal response ...
2024/09/20 19:34:05 Ready to write response ...
2024/09/20 19:34:19 Ready to marshal response ...
2024/09/20 19:34:19 Ready to write response ...
2024/09/20 19:34:41 Ready to marshal response ...
2024/09/20 19:34:41 Ready to write response ...
2024/09/20 19:34:41 Ready to marshal response ...
2024/09/20 19:34:41 Ready to write response ...
2024/09/20 19:34:50 Ready to marshal response ...
2024/09/20 19:34:50 Ready to write response ...
==> kernel <==
19:35:07 up 3:17, 0 users, load average: 0.85, 0.87, 1.46
Linux addons-711398 5.15.0-1070-aws #76~20.04.1-Ubuntu SMP Mon Sep 2 12:20:48 UTC 2024 aarch64 aarch64 aarch64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [c13c74eade15] <==
I0920 19:25:41.157756 1 handler.go:286] Adding GroupVersion flow.volcano.sh v1alpha1 to ResourceManager
I0920 19:25:41.190607 1 handler.go:286] Adding GroupVersion flow.volcano.sh v1alpha1 to ResourceManager
I0920 19:25:41.259618 1 handler.go:286] Adding GroupVersion flow.volcano.sh v1alpha1 to ResourceManager
W0920 19:25:41.376496 1 cacher.go:171] Terminating all watchers from cacher commands.bus.volcano.sh
W0920 19:25:41.894327 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0920 19:25:41.905403 1 cacher.go:171] Terminating all watchers from cacher podgroups.scheduling.volcano.sh
W0920 19:25:41.906749 1 cacher.go:171] Terminating all watchers from cacher numatopologies.nodeinfo.volcano.sh
W0920 19:25:41.997080 1 cacher.go:171] Terminating all watchers from cacher queues.scheduling.volcano.sh
W0920 19:25:42.260287 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0920 19:25:42.596745 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0920 19:34:13.042365 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0920 19:34:34.390560 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0920 19:34:34.390608 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0920 19:34:34.414927 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0920 19:34:34.414988 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0920 19:34:34.438134 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0920 19:34:34.438185 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0920 19:34:34.453428 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0920 19:34:34.453481 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0920 19:34:34.569737 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0920 19:34:34.569781 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0920 19:34:35.439002 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W0920 19:34:35.570435 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
W0920 19:34:35.587110 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
E0920 19:35:06.406181 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
==> kube-controller-manager [cc9b700d2590] <==
W0920 19:34:38.726082 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 19:34:38.726126 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 19:34:39.192866 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 19:34:39.192911 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 19:34:42.951835 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 19:34:42.951877 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 19:34:43.652739 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 19:34:43.652793 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 19:34:44.492378 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 19:34:44.492425 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0920 19:34:51.042636 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="local-path-storage/local-path-provisioner-86d989889c" duration="4.521µs"
I0920 19:34:51.658842 1 shared_informer.go:313] Waiting for caches to sync for resource quota
I0920 19:34:51.658881 1 shared_informer.go:320] Caches are synced for resource quota
I0920 19:34:51.895321 1 shared_informer.go:313] Waiting for caches to sync for garbage collector
I0920 19:34:51.896039 1 shared_informer.go:320] Caches are synced for garbage collector
I0920 19:34:52.312716 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-711398"
W0920 19:34:54.234836 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 19:34:54.234879 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 19:34:54.745117 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 19:34:54.745161 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 19:34:56.723805 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 19:34:56.723848 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0920 19:35:05.760448 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="4.48µs"
W0920 19:35:06.879455 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 19:35:06.879500 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
==> kube-proxy [060c8c64d422] <==
I0920 19:22:23.348320 1 server_linux.go:66] "Using iptables proxy"
I0920 19:22:23.515835 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0920 19:22:23.515917 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0920 19:22:23.545966 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0920 19:22:23.546030 1 server_linux.go:169] "Using iptables Proxier"
I0920 19:22:23.548675 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0920 19:22:23.549066 1 server.go:483] "Version info" version="v1.31.1"
I0920 19:22:23.549097 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0920 19:22:23.550901 1 config.go:199] "Starting service config controller"
I0920 19:22:23.550929 1 shared_informer.go:313] Waiting for caches to sync for service config
I0920 19:22:23.550978 1 config.go:105] "Starting endpoint slice config controller"
I0920 19:22:23.550983 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0920 19:22:23.553928 1 config.go:328] "Starting node config controller"
I0920 19:22:23.553943 1 shared_informer.go:313] Waiting for caches to sync for node config
I0920 19:22:23.651908 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0920 19:22:23.651997 1 shared_informer.go:320] Caches are synced for service config
I0920 19:22:23.654148 1 shared_informer.go:320] Caches are synced for node config
==> kube-scheduler [23d47197484d] <==
W0920 19:22:14.919140 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0920 19:22:14.919226 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 19:22:14.919399 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0920 19:22:14.919479 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 19:22:14.919604 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0920 19:22:14.919706 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 19:22:14.919943 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0920 19:22:14.920030 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0920 19:22:14.920284 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0920 19:22:14.920771 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 19:22:14.920364 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0920 19:22:14.921227 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0920 19:22:14.920403 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0920 19:22:14.921388 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 19:22:14.920448 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0920 19:22:14.921675 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0920 19:22:14.920503 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0920 19:22:14.922143 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0920 19:22:14.921801 1 reflector.go:561] runtime/asm_arm64.s:1222: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
W0920 19:22:14.920563 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
W0920 19:22:14.920633 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0920 19:22:14.922589 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
E0920 19:22:14.922656 1 reflector.go:158] "Unhandled Error" err="runtime/asm_arm64.s:1222: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
E0920 19:22:14.922752 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
I0920 19:22:16.307911 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 20 19:34:54 addons-711398 kubelet[2326]: E0920 19:34:54.508572 2326 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"gadget\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=gadget pod=gadget-hp9tl_gadget(53daf043-7f53-406d-ac9d-0815258bc4b1)\"" pod="gadget/gadget-hp9tl" podUID="53daf043-7f53-406d-ac9d-0815258bc4b1"
Sep 20 19:34:56 addons-711398 kubelet[2326]: I0920 19:34:56.517960 2326 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02081004-efea-460d-8eec-c014cf579521" path="/var/lib/kubelet/pods/02081004-efea-460d-8eec-c014cf579521/volumes"
Sep 20 19:34:59 addons-711398 kubelet[2326]: E0920 19:34:59.510125 2326 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="f8da1040-f0d0-4fec-9895-578b53a3b266"
Sep 20 19:35:03 addons-711398 kubelet[2326]: E0920 19:35:03.509938 2326 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox\\\"\"" pod="default/registry-test" podUID="66a65f15-36bd-442f-9ee9-7d762e51c91c"
Sep 20 19:35:05 addons-711398 kubelet[2326]: I0920 19:35:05.367133 2326 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/66a65f15-36bd-442f-9ee9-7d762e51c91c-gcp-creds\") pod \"66a65f15-36bd-442f-9ee9-7d762e51c91c\" (UID: \"66a65f15-36bd-442f-9ee9-7d762e51c91c\") "
Sep 20 19:35:05 addons-711398 kubelet[2326]: I0920 19:35:05.367199 2326 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59kxw\" (UniqueName: \"kubernetes.io/projected/66a65f15-36bd-442f-9ee9-7d762e51c91c-kube-api-access-59kxw\") pod \"66a65f15-36bd-442f-9ee9-7d762e51c91c\" (UID: \"66a65f15-36bd-442f-9ee9-7d762e51c91c\") "
Sep 20 19:35:05 addons-711398 kubelet[2326]: I0920 19:35:05.368279 2326 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66a65f15-36bd-442f-9ee9-7d762e51c91c-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "66a65f15-36bd-442f-9ee9-7d762e51c91c" (UID: "66a65f15-36bd-442f-9ee9-7d762e51c91c"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Sep 20 19:35:05 addons-711398 kubelet[2326]: I0920 19:35:05.372650 2326 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66a65f15-36bd-442f-9ee9-7d762e51c91c-kube-api-access-59kxw" (OuterVolumeSpecName: "kube-api-access-59kxw") pod "66a65f15-36bd-442f-9ee9-7d762e51c91c" (UID: "66a65f15-36bd-442f-9ee9-7d762e51c91c"). InnerVolumeSpecName "kube-api-access-59kxw". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 20 19:35:05 addons-711398 kubelet[2326]: I0920 19:35:05.468853 2326 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/66a65f15-36bd-442f-9ee9-7d762e51c91c-gcp-creds\") on node \"addons-711398\" DevicePath \"\""
Sep 20 19:35:05 addons-711398 kubelet[2326]: I0920 19:35:05.468892 2326 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-59kxw\" (UniqueName: \"kubernetes.io/projected/66a65f15-36bd-442f-9ee9-7d762e51c91c-kube-api-access-59kxw\") on node \"addons-711398\" DevicePath \"\""
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.275717 2326 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lftr\" (UniqueName: \"kubernetes.io/projected/d2e45ba0-4b0a-4648-a233-1dfc5982c286-kube-api-access-2lftr\") pod \"d2e45ba0-4b0a-4648-a233-1dfc5982c286\" (UID: \"d2e45ba0-4b0a-4648-a233-1dfc5982c286\") "
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.279817 2326 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e45ba0-4b0a-4648-a233-1dfc5982c286-kube-api-access-2lftr" (OuterVolumeSpecName: "kube-api-access-2lftr") pod "d2e45ba0-4b0a-4648-a233-1dfc5982c286" (UID: "d2e45ba0-4b0a-4648-a233-1dfc5982c286"). InnerVolumeSpecName "kube-api-access-2lftr". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.376488 2326 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx75b\" (UniqueName: \"kubernetes.io/projected/4fbca207-de93-4adb-baa8-2219f829573b-kube-api-access-wx75b\") pod \"4fbca207-de93-4adb-baa8-2219f829573b\" (UID: \"4fbca207-de93-4adb-baa8-2219f829573b\") "
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.376622 2326 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-2lftr\" (UniqueName: \"kubernetes.io/projected/d2e45ba0-4b0a-4648-a233-1dfc5982c286-kube-api-access-2lftr\") on node \"addons-711398\" DevicePath \"\""
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.381046 2326 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fbca207-de93-4adb-baa8-2219f829573b-kube-api-access-wx75b" (OuterVolumeSpecName: "kube-api-access-wx75b") pod "4fbca207-de93-4adb-baa8-2219f829573b" (UID: "4fbca207-de93-4adb-baa8-2219f829573b"). InnerVolumeSpecName "kube-api-access-wx75b". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.477101 2326 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-wx75b\" (UniqueName: \"kubernetes.io/projected/4fbca207-de93-4adb-baa8-2219f829573b-kube-api-access-wx75b\") on node \"addons-711398\" DevicePath \"\""
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.526788 2326 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66a65f15-36bd-442f-9ee9-7d762e51c91c" path="/var/lib/kubelet/pods/66a65f15-36bd-442f-9ee9-7d762e51c91c/volumes"
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.709052 2326 scope.go:117] "RemoveContainer" containerID="c8ddd2e4784afd8f3ed04ab0294c18ddc8d819776f9b2f7cc18c6fb4fd526e2c"
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.760198 2326 scope.go:117] "RemoveContainer" containerID="c8ddd2e4784afd8f3ed04ab0294c18ddc8d819776f9b2f7cc18c6fb4fd526e2c"
Sep 20 19:35:06 addons-711398 kubelet[2326]: E0920 19:35:06.761419 2326 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: c8ddd2e4784afd8f3ed04ab0294c18ddc8d819776f9b2f7cc18c6fb4fd526e2c" containerID="c8ddd2e4784afd8f3ed04ab0294c18ddc8d819776f9b2f7cc18c6fb4fd526e2c"
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.761634 2326 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"c8ddd2e4784afd8f3ed04ab0294c18ddc8d819776f9b2f7cc18c6fb4fd526e2c"} err="failed to get container status \"c8ddd2e4784afd8f3ed04ab0294c18ddc8d819776f9b2f7cc18c6fb4fd526e2c\": rpc error: code = Unknown desc = Error response from daemon: No such container: c8ddd2e4784afd8f3ed04ab0294c18ddc8d819776f9b2f7cc18c6fb4fd526e2c"
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.761743 2326 scope.go:117] "RemoveContainer" containerID="f8330d199113ddd1713739370a5f6157c8a097384c87635281582db2e6df1481"
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.794344 2326 scope.go:117] "RemoveContainer" containerID="f8330d199113ddd1713739370a5f6157c8a097384c87635281582db2e6df1481"
Sep 20 19:35:06 addons-711398 kubelet[2326]: E0920 19:35:06.796536 2326 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: f8330d199113ddd1713739370a5f6157c8a097384c87635281582db2e6df1481" containerID="f8330d199113ddd1713739370a5f6157c8a097384c87635281582db2e6df1481"
Sep 20 19:35:06 addons-711398 kubelet[2326]: I0920 19:35:06.796614 2326 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"f8330d199113ddd1713739370a5f6157c8a097384c87635281582db2e6df1481"} err="failed to get container status \"f8330d199113ddd1713739370a5f6157c8a097384c87635281582db2e6df1481\": rpc error: code = Unknown desc = Error response from daemon: No such container: f8330d199113ddd1713739370a5f6157c8a097384c87635281582db2e6df1481"
==> storage-provisioner [dcbbde731377] <==
I0920 19:22:28.186265 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0920 19:22:28.206714 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0920 19:22:28.206758 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0920 19:22:28.224439 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0920 19:22:28.224690 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"406391ed-d0f5-42dc-b149-3c882314ab08", APIVersion:"v1", ResourceVersion:"531", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-711398_6bec32ad-bee2-4c3b-8880-2b31f1cab225 became leader
I0920 19:22:28.225338 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-711398_6bec32ad-bee2-4c3b-8880-2b31f1cab225!
I0920 19:22:28.328206 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-711398_6bec32ad-bee2-4c3b-8880-2b31f1cab225!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-arm64 status --format={{.APIServer}} -p addons-711398 -n addons-711398
helpers_test.go:261: (dbg) Run: kubectl --context addons-711398 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox ingress-nginx-admission-create-7mls9 ingress-nginx-admission-patch-n2285
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-711398 describe pod busybox ingress-nginx-admission-create-7mls9 ingress-nginx-admission-patch-n2285
helpers_test.go:277: (dbg) Non-zero exit: kubectl --context addons-711398 describe pod busybox ingress-nginx-admission-create-7mls9 ingress-nginx-admission-patch-n2285: exit status 1 (103.95304ms)
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-711398/192.168.49.2
Start Time: Fri, 20 Sep 2024 19:25:50 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.27
IPs:
IP: 10.244.0.27
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-n2g9z (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-n2g9z:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m18s default-scheduler Successfully assigned default/busybox to addons-711398
Normal Pulling 7m54s (x4 over 9m17s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m54s (x4 over 9m17s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m54s (x4 over 9m17s) kubelet Error: ErrImagePull
Warning Failed 7m42s (x6 over 9m16s) kubelet Error: ImagePullBackOff
Normal BackOff 4m7s (x22 over 9m16s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
** stderr **
Error from server (NotFound): pods "ingress-nginx-admission-create-7mls9" not found
Error from server (NotFound): pods "ingress-nginx-admission-patch-n2285" not found
** /stderr **
helpers_test.go:279: kubectl --context addons-711398 describe pod busybox ingress-nginx-admission-create-7mls9 ingress-nginx-admission-patch-n2285: exit status 1
--- FAIL: TestAddons/parallel/Registry (74.48s)