=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:332: registry stabilized in 1.888178ms
addons_test.go:334: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-6fb4cdfc84-lbgl8" [243f93fd-9335-4ba8-a33c-d9ed0affd92f] Running
addons_test.go:334: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.003137047s
addons_test.go:337: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-vxpf8" [6d3d6e6b-090e-4033-9117-635faf64b5bf] Running
addons_test.go:337: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.003328325s
addons_test.go:342: (dbg) Run: kubectl --context addons-001161 delete po -l run=registry-test --now
addons_test.go:347: (dbg) Run: kubectl --context addons-001161 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:347: (dbg) Non-zero exit: kubectl --context addons-001161 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.071656568s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:349: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-001161 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:353: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:361: (dbg) Run: out/minikube-linux-amd64 -p addons-001161 ip
2024/08/28 17:05:05 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:390: (dbg) Run: out/minikube-linux-amd64 -p addons-001161 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-001161
helpers_test.go:235: (dbg) docker inspect addons-001161:
-- stdout --
[
{
"Id": "31fae27b33dc7ecf2a97bc05ecf350845d4075ba75c02994500da1fc99fd5294",
"Created": "2024-08-28T16:52:07.396042766Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 19049,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-08-28T16:52:07.524544922Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:33319d96a2f78fe466b6d8cbd88671515fca2b1eded3ce0b5f6d545b670a78ac",
"ResolvConfPath": "/var/lib/docker/containers/31fae27b33dc7ecf2a97bc05ecf350845d4075ba75c02994500da1fc99fd5294/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/31fae27b33dc7ecf2a97bc05ecf350845d4075ba75c02994500da1fc99fd5294/hostname",
"HostsPath": "/var/lib/docker/containers/31fae27b33dc7ecf2a97bc05ecf350845d4075ba75c02994500da1fc99fd5294/hosts",
"LogPath": "/var/lib/docker/containers/31fae27b33dc7ecf2a97bc05ecf350845d4075ba75c02994500da1fc99fd5294/31fae27b33dc7ecf2a97bc05ecf350845d4075ba75c02994500da1fc99fd5294-json.log",
"Name": "/addons-001161",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-001161:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-001161",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/8b6fb6d10384824a02064181e491b96c49d3b6434a16c7c148b2537fe5ad04d0-init/diff:/var/lib/docker/overlay2/1e993f01f09edd4a15c10f1631ac1eafc9f77a8fb55b51a6a7703d841f0383cb/diff",
"MergedDir": "/var/lib/docker/overlay2/8b6fb6d10384824a02064181e491b96c49d3b6434a16c7c148b2537fe5ad04d0/merged",
"UpperDir": "/var/lib/docker/overlay2/8b6fb6d10384824a02064181e491b96c49d3b6434a16c7c148b2537fe5ad04d0/diff",
"WorkDir": "/var/lib/docker/overlay2/8b6fb6d10384824a02064181e491b96c49d3b6434a16c7c148b2537fe5ad04d0/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-001161",
"Source": "/var/lib/docker/volumes/addons-001161/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-001161",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-001161",
"name.minikube.sigs.k8s.io": "addons-001161",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "313fef41e40ed38c505ccbc71c2b2fd51cd6cc357adbac80c33ddbbbdcd6ee02",
"SandboxKey": "/var/run/docker/netns/313fef41e40e",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-001161": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "533a02b9040e0370fb27139c47d4907663144582e4549c5acc8e883802360202",
"EndpointID": "ac315ef541f576ec249c5ebae8638f0c3b541ea04e42f968bb12921ae2087730",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-001161",
"31fae27b33dc"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-001161 -n addons-001161
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-001161 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-docker-307047 | download-docker-307047 | jenkins | v1.33.1 | 28 Aug 24 16:51 UTC | 28 Aug 24 16:51 UTC |
| start | --download-only -p | binary-mirror-380970 | jenkins | v1.33.1 | 28 Aug 24 16:51 UTC | |
| | binary-mirror-380970 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:33503 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-380970 | binary-mirror-380970 | jenkins | v1.33.1 | 28 Aug 24 16:51 UTC | 28 Aug 24 16:51 UTC |
| addons | enable dashboard -p | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 16:51 UTC | |
| | addons-001161 | | | | | |
| addons | disable dashboard -p | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 16:51 UTC | |
| | addons-001161 | | | | | |
| start | -p addons-001161 --wait=true | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 16:51 UTC | 28 Aug 24 16:55 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | addons-001161 addons disable | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 16:55 UTC | 28 Aug 24 16:55 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | addons-001161 addons | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-001161 ssh cat | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | /opt/local-path-provisioner/pvc-26a8e827-e6a9-4d7b-a824-2c6c759b92e6_default_test-pvc/file1 | | | | | |
| addons | addons-001161 addons disable | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-001161 addons disable | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-001161 addons disable | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | -p addons-001161 | | | | | |
| addons | disable inspektor-gadget -p | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | addons-001161 | | | | | |
| addons | disable cloud-spanner -p | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | addons-001161 | | | | | |
| addons | enable headlamp | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | -p addons-001161 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-001161 addons disable | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ssh | addons-001161 ssh curl -s | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-001161 ip | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| addons | addons-001161 addons disable | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-001161 addons disable | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| addons | addons-001161 addons | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-001161 addons | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:04 UTC | 28 Aug 24 17:04 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-001161 ip | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:05 UTC | 28 Aug 24 17:05 UTC |
| addons | addons-001161 addons disable | addons-001161 | jenkins | v1.33.1 | 28 Aug 24 17:05 UTC | 28 Aug 24 17:05 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/08/28 16:51:46
Running on machine: ubuntu-20-agent
Binary: Built with gc go1.22.5 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0828 16:51:46.127569 18312 out.go:345] Setting OutFile to fd 1 ...
I0828 16:51:46.127831 18312 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0828 16:51:46.127842 18312 out.go:358] Setting ErrFile to fd 2...
I0828 16:51:46.127847 18312 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0828 16:51:46.128081 18312 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19529-10133/.minikube/bin
I0828 16:51:46.128755 18312 out.go:352] Setting JSON to false
I0828 16:51:46.129553 18312 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent","uptime":2052,"bootTime":1724861854,"procs":170,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1067-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0828 16:51:46.129615 18312 start.go:139] virtualization: kvm guest
I0828 16:51:46.131693 18312 out.go:177] * [addons-001161] minikube v1.33.1 on Ubuntu 20.04 (kvm/amd64)
I0828 16:51:46.132966 18312 notify.go:220] Checking for updates...
I0828 16:51:46.132998 18312 out.go:177] - MINIKUBE_LOCATION=19529
I0828 16:51:46.134300 18312 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0828 16:51:46.135515 18312 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19529-10133/kubeconfig
I0828 16:51:46.136792 18312 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19529-10133/.minikube
I0828 16:51:46.138090 18312 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0828 16:51:46.139329 18312 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0828 16:51:46.140724 18312 driver.go:392] Setting default libvirt URI to qemu:///system
I0828 16:51:46.162038 18312 docker.go:123] docker version: linux-27.2.0:Docker Engine - Community
I0828 16:51:46.162155 18312 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0828 16:51:46.207695 18312 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-08-28 16:51:46.199022182 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1067-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647935488 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent Labels:[] ExperimentalBuild:false ServerVersion:27.2.0 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:472731909fa34bd7bc9c087e4c27943f9835f111 Expected:472731909fa34bd7bc9c087e4c27943f9835f111} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerError
s:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0828 16:51:46.207805 18312 docker.go:307] overlay module found
I0828 16:51:46.209708 18312 out.go:177] * Using the docker driver based on user configuration
I0828 16:51:46.210875 18312 start.go:297] selected driver: docker
I0828 16:51:46.210890 18312 start.go:901] validating driver "docker" against <nil>
I0828 16:51:46.210901 18312 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0828 16:51:46.211693 18312 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0828 16:51:46.255023 18312 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-08-28 16:51:46.24676043 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1067-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86_
64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647935488 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent Labels:[] ExperimentalBuild:false ServerVersion:27.2.0 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:472731909fa34bd7bc9c087e4c27943f9835f111 Expected:472731909fa34bd7bc9c087e4c27943f9835f111} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErrors
:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0828 16:51:46.255182 18312 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0828 16:51:46.255371 18312 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0828 16:51:46.257086 18312 out.go:177] * Using Docker driver with root privileges
I0828 16:51:46.258470 18312 cni.go:84] Creating CNI manager for ""
I0828 16:51:46.258489 18312 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0828 16:51:46.258499 18312 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0828 16:51:46.258558 18312 start.go:340] cluster config:
{Name:addons-001161 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-001161 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0828 16:51:46.259841 18312 out.go:177] * Starting "addons-001161" primary control-plane node in "addons-001161" cluster
I0828 16:51:46.261106 18312 cache.go:121] Beginning downloading kic base image for docker with docker
I0828 16:51:46.262412 18312 out.go:177] * Pulling base image v0.0.44-1724775115-19521 ...
I0828 16:51:46.263519 18312 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0828 16:51:46.263545 18312 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19529-10133/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4
I0828 16:51:46.263553 18312 cache.go:56] Caching tarball of preloaded images
I0828 16:51:46.263625 18312 preload.go:172] Found /home/jenkins/minikube-integration/19529-10133/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0828 16:51:46.263639 18312 cache.go:59] Finished verifying existence of preloaded tar for v1.31.0 on docker
I0828 16:51:46.263623 18312 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce in local docker daemon
I0828 16:51:46.263983 18312 profile.go:143] Saving config to /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/config.json ...
I0828 16:51:46.264009 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/config.json: {Name:mk0ea066effc3a0807efa2f8a19e482b7787dbbb Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:51:46.280308 18312 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce to local cache
I0828 16:51:46.280416 18312 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce in local cache directory
I0828 16:51:46.280432 18312 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce in local cache directory, skipping pull
I0828 16:51:46.280436 18312 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce exists in cache, skipping pull
I0828 16:51:46.280443 18312 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce as a tarball
I0828 16:51:46.280450 18312 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce from local cache
I0828 16:51:58.233751 18312 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce from cached tarball
I0828 16:51:58.233800 18312 cache.go:194] Successfully downloaded all kic artifacts
I0828 16:51:58.233831 18312 start.go:360] acquireMachinesLock for addons-001161: {Name:mk6c6aab2c8f2d19f1935e90285021f75273bd01 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0828 16:51:58.233917 18312 start.go:364] duration metric: took 68.58µs to acquireMachinesLock for "addons-001161"
I0828 16:51:58.233941 18312 start.go:93] Provisioning new machine with config: &{Name:addons-001161 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-001161 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0828 16:51:58.234022 18312 start.go:125] createHost starting for "" (driver="docker")
I0828 16:51:58.235929 18312 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0828 16:51:58.236139 18312 start.go:159] libmachine.API.Create for "addons-001161" (driver="docker")
I0828 16:51:58.236168 18312 client.go:168] LocalClient.Create starting
I0828 16:51:58.236279 18312 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19529-10133/.minikube/certs/ca.pem
I0828 16:51:58.330114 18312 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19529-10133/.minikube/certs/cert.pem
I0828 16:51:58.419468 18312 cli_runner.go:164] Run: docker network inspect addons-001161 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0828 16:51:58.434937 18312 cli_runner.go:211] docker network inspect addons-001161 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0828 16:51:58.435019 18312 network_create.go:284] running [docker network inspect addons-001161] to gather additional debugging logs...
I0828 16:51:58.435040 18312 cli_runner.go:164] Run: docker network inspect addons-001161
W0828 16:51:58.449359 18312 cli_runner.go:211] docker network inspect addons-001161 returned with exit code 1
I0828 16:51:58.449386 18312 network_create.go:287] error running [docker network inspect addons-001161]: docker network inspect addons-001161: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-001161 not found
I0828 16:51:58.449410 18312 network_create.go:289] output of [docker network inspect addons-001161]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-001161 not found
** /stderr **
I0828 16:51:58.449521 18312 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0828 16:51:58.464207 18312 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001e704f0}
I0828 16:51:58.464271 18312 network_create.go:124] attempt to create docker network addons-001161 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0828 16:51:58.464308 18312 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-001161 addons-001161
I0828 16:51:58.521239 18312 network_create.go:108] docker network addons-001161 192.168.49.0/24 created
I0828 16:51:58.521271 18312 kic.go:121] calculated static IP "192.168.49.2" for the "addons-001161" container
I0828 16:51:58.521336 18312 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0828 16:51:58.535686 18312 cli_runner.go:164] Run: docker volume create addons-001161 --label name.minikube.sigs.k8s.io=addons-001161 --label created_by.minikube.sigs.k8s.io=true
I0828 16:51:58.551406 18312 oci.go:103] Successfully created a docker volume addons-001161
I0828 16:51:58.551505 18312 cli_runner.go:164] Run: docker run --rm --name addons-001161-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-001161 --entrypoint /usr/bin/test -v addons-001161:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce -d /var/lib
I0828 16:52:03.440552 18312 cli_runner.go:217] Completed: docker run --rm --name addons-001161-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-001161 --entrypoint /usr/bin/test -v addons-001161:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce -d /var/lib: (4.889009777s)
I0828 16:52:03.440577 18312 oci.go:107] Successfully prepared a docker volume addons-001161
I0828 16:52:03.440600 18312 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0828 16:52:03.440622 18312 kic.go:194] Starting extracting preloaded images to volume ...
I0828 16:52:03.440682 18312 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19529-10133/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-001161:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce -I lz4 -xf /preloaded.tar -C /extractDir
I0828 16:52:07.332817 18312 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19529-10133/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-001161:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce -I lz4 -xf /preloaded.tar -C /extractDir: (3.892099986s)
I0828 16:52:07.332846 18312 kic.go:203] duration metric: took 3.892221702s to extract preloaded images to volume ...
W0828 16:52:07.332980 18312 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0828 16:52:07.333086 18312 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0828 16:52:07.382562 18312 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-001161 --name addons-001161 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-001161 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-001161 --network addons-001161 --ip 192.168.49.2 --volume addons-001161:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce
I0828 16:52:07.680072 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Running}}
I0828 16:52:07.697957 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:07.715804 18312 cli_runner.go:164] Run: docker exec addons-001161 stat /var/lib/dpkg/alternatives/iptables
I0828 16:52:07.756748 18312 oci.go:144] the created container "addons-001161" has a running status.
I0828 16:52:07.756780 18312 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa...
I0828 16:52:07.931603 18312 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0828 16:52:07.954251 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:07.978036 18312 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0828 16:52:07.978061 18312 kic_runner.go:114] Args: [docker exec --privileged addons-001161 chown docker:docker /home/docker/.ssh/authorized_keys]
I0828 16:52:08.039972 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:08.058237 18312 machine.go:93] provisionDockerMachine start ...
I0828 16:52:08.058317 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:08.077055 18312 main.go:141] libmachine: Using SSH client type: native
I0828 16:52:08.077440 18312 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0828 16:52:08.077456 18312 main.go:141] libmachine: About to run SSH command:
hostname
I0828 16:52:08.251731 18312 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-001161
I0828 16:52:08.251763 18312 ubuntu.go:169] provisioning hostname "addons-001161"
I0828 16:52:08.251819 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:08.270353 18312 main.go:141] libmachine: Using SSH client type: native
I0828 16:52:08.270521 18312 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0828 16:52:08.270533 18312 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-001161 && echo "addons-001161" | sudo tee /etc/hostname
I0828 16:52:08.405872 18312 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-001161
I0828 16:52:08.405949 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:08.421661 18312 main.go:141] libmachine: Using SSH client type: native
I0828 16:52:08.421881 18312 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0828 16:52:08.421906 18312 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-001161' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-001161/g' /etc/hosts;
else
echo '127.0.1.1 addons-001161' | sudo tee -a /etc/hosts;
fi
fi
I0828 16:52:08.543864 18312 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0828 16:52:08.543893 18312 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19529-10133/.minikube CaCertPath:/home/jenkins/minikube-integration/19529-10133/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19529-10133/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19529-10133/.minikube}
I0828 16:52:08.543915 18312 ubuntu.go:177] setting up certificates
I0828 16:52:08.543924 18312 provision.go:84] configureAuth start
I0828 16:52:08.543976 18312 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-001161
I0828 16:52:08.560800 18312 provision.go:143] copyHostCerts
I0828 16:52:08.560875 18312 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19529-10133/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19529-10133/.minikube/ca.pem (1082 bytes)
I0828 16:52:08.560982 18312 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19529-10133/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19529-10133/.minikube/cert.pem (1123 bytes)
I0828 16:52:08.561038 18312 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19529-10133/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19529-10133/.minikube/key.pem (1679 bytes)
I0828 16:52:08.561084 18312 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19529-10133/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19529-10133/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19529-10133/.minikube/certs/ca-key.pem org=jenkins.addons-001161 san=[127.0.0.1 192.168.49.2 addons-001161 localhost minikube]
I0828 16:52:08.776260 18312 provision.go:177] copyRemoteCerts
I0828 16:52:08.776309 18312 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0828 16:52:08.776345 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:08.792166 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:08.879929 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0828 16:52:08.900050 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I0828 16:52:08.919517 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0828 16:52:08.938960 18312 provision.go:87] duration metric: took 395.024976ms to configureAuth
I0828 16:52:08.938987 18312 ubuntu.go:193] setting minikube options for container-runtime
I0828 16:52:08.939154 18312 config.go:182] Loaded profile config "addons-001161": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0828 16:52:08.939203 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:08.955013 18312 main.go:141] libmachine: Using SSH client type: native
I0828 16:52:08.955173 18312 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0828 16:52:08.955186 18312 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0828 16:52:09.076151 18312 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0828 16:52:09.076174 18312 ubuntu.go:71] root file system type: overlay
I0828 16:52:09.076318 18312 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0828 16:52:09.076372 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:09.091751 18312 main.go:141] libmachine: Using SSH client type: native
I0828 16:52:09.091910 18312 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0828 16:52:09.091966 18312 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0828 16:52:09.221534 18312 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0828 16:52:09.221608 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:09.237290 18312 main.go:141] libmachine: Using SSH client type: native
I0828 16:52:09.237490 18312 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0828 16:52:09.237518 18312 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0828 16:52:09.888498 18312 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-08-12 11:48:57.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-08-28 16:52:09.219222398 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0828 16:52:09.888538 18312 machine.go:96] duration metric: took 1.830281507s to provisionDockerMachine
I0828 16:52:09.888553 18312 client.go:171] duration metric: took 11.652376361s to LocalClient.Create
I0828 16:52:09.888574 18312 start.go:167] duration metric: took 11.652434929s to libmachine.API.Create "addons-001161"
I0828 16:52:09.888587 18312 start.go:293] postStartSetup for "addons-001161" (driver="docker")
I0828 16:52:09.888599 18312 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0828 16:52:09.888662 18312 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0828 16:52:09.888710 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:09.904131 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:09.992138 18312 ssh_runner.go:195] Run: cat /etc/os-release
I0828 16:52:09.994976 18312 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0828 16:52:09.995009 18312 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0828 16:52:09.995020 18312 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0828 16:52:09.995028 18312 info.go:137] Remote host: Ubuntu 22.04.4 LTS
I0828 16:52:09.995038 18312 filesync.go:126] Scanning /home/jenkins/minikube-integration/19529-10133/.minikube/addons for local assets ...
I0828 16:52:09.995090 18312 filesync.go:126] Scanning /home/jenkins/minikube-integration/19529-10133/.minikube/files for local assets ...
I0828 16:52:09.995112 18312 start.go:296] duration metric: took 106.519686ms for postStartSetup
I0828 16:52:09.995346 18312 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-001161
I0828 16:52:10.011438 18312 profile.go:143] Saving config to /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/config.json ...
I0828 16:52:10.011663 18312 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0828 16:52:10.011707 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:10.027202 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:10.112600 18312 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0828 16:52:10.116428 18312 start.go:128] duration metric: took 11.882390394s to createHost
I0828 16:52:10.116445 18312 start.go:83] releasing machines lock for "addons-001161", held for 11.882517711s
I0828 16:52:10.116506 18312 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-001161
I0828 16:52:10.132096 18312 ssh_runner.go:195] Run: cat /version.json
I0828 16:52:10.132139 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:10.132202 18312 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0828 16:52:10.132283 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:10.148705 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:10.149252 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:10.235481 18312 ssh_runner.go:195] Run: systemctl --version
I0828 16:52:10.239365 18312 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0828 16:52:10.307288 18312 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0828 16:52:10.329479 18312 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0828 16:52:10.329539 18312 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0828 16:52:10.353026 18312 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0828 16:52:10.353051 18312 start.go:495] detecting cgroup driver to use...
I0828 16:52:10.353078 18312 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0828 16:52:10.353171 18312 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0828 16:52:10.366763 18312 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0828 16:52:10.375022 18312 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0828 16:52:10.383478 18312 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0828 16:52:10.383531 18312 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0828 16:52:10.392085 18312 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0828 16:52:10.400799 18312 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0828 16:52:10.409232 18312 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0828 16:52:10.417617 18312 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0828 16:52:10.425465 18312 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0828 16:52:10.434141 18312 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0828 16:52:10.442743 18312 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0828 16:52:10.451381 18312 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0828 16:52:10.458461 18312 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0828 16:52:10.465566 18312 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0828 16:52:10.540462 18312 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0828 16:52:10.624791 18312 start.go:495] detecting cgroup driver to use...
I0828 16:52:10.624834 18312 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0828 16:52:10.624872 18312 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0828 16:52:10.636594 18312 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0828 16:52:10.636669 18312 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0828 16:52:10.647208 18312 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0828 16:52:10.662233 18312 ssh_runner.go:195] Run: which cri-dockerd
I0828 16:52:10.665782 18312 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0828 16:52:10.674066 18312 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0828 16:52:10.690793 18312 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0828 16:52:10.773096 18312 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0828 16:52:10.869667 18312 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0828 16:52:10.869795 18312 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0828 16:52:10.885403 18312 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0828 16:52:10.962711 18312 ssh_runner.go:195] Run: sudo systemctl restart docker
I0828 16:52:11.209372 18312 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0828 16:52:11.219716 18312 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0828 16:52:11.229853 18312 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0828 16:52:11.303537 18312 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0828 16:52:11.385212 18312 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0828 16:52:11.456975 18312 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0828 16:52:11.468505 18312 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0828 16:52:11.477843 18312 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0828 16:52:11.553609 18312 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0828 16:52:11.610645 18312 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0828 16:52:11.610718 18312 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0828 16:52:11.614110 18312 start.go:563] Will wait 60s for crictl version
I0828 16:52:11.614164 18312 ssh_runner.go:195] Run: which crictl
I0828 16:52:11.617000 18312 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0828 16:52:11.646656 18312 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.1.2
RuntimeApiVersion: v1
I0828 16:52:11.646725 18312 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0828 16:52:11.669051 18312 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0828 16:52:11.691913 18312 out.go:235] * Preparing Kubernetes v1.31.0 on Docker 27.1.2 ...
I0828 16:52:11.691999 18312 cli_runner.go:164] Run: docker network inspect addons-001161 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0828 16:52:11.708061 18312 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0828 16:52:11.711391 18312 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0828 16:52:11.720895 18312 kubeadm.go:883] updating cluster {Name:addons-001161 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-001161 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0828 16:52:11.720995 18312 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0828 16:52:11.721043 18312 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0828 16:52:11.738872 18312 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.0
registry.k8s.io/kube-controller-manager:v1.31.0
registry.k8s.io/kube-scheduler:v1.31.0
registry.k8s.io/kube-proxy:v1.31.0
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
registry.k8s.io/coredns/coredns:v1.11.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0828 16:52:11.738891 18312 docker.go:615] Images already preloaded, skipping extraction
I0828 16:52:11.738935 18312 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0828 16:52:11.756194 18312 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.0
registry.k8s.io/kube-controller-manager:v1.31.0
registry.k8s.io/kube-scheduler:v1.31.0
registry.k8s.io/kube-proxy:v1.31.0
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
registry.k8s.io/coredns/coredns:v1.11.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0828 16:52:11.756215 18312 cache_images.go:84] Images are preloaded, skipping loading
I0828 16:52:11.756236 18312 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.0 docker true true} ...
I0828 16:52:11.756360 18312 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.0/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-001161 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.0 ClusterName:addons-001161 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0828 16:52:11.756422 18312 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0828 16:52:11.798771 18312 cni.go:84] Creating CNI manager for ""
I0828 16:52:11.798800 18312 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0828 16:52:11.798810 18312 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0828 16:52:11.798829 18312 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.0 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-001161 NodeName:addons-001161 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0828 16:52:11.798958 18312 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-001161"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.0
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0828 16:52:11.799014 18312 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.0
I0828 16:52:11.806936 18312 binaries.go:44] Found k8s binaries, skipping transfer
I0828 16:52:11.806998 18312 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0828 16:52:11.814678 18312 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0828 16:52:11.830013 18312 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0828 16:52:11.845545 18312 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0828 16:52:11.860674 18312 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0828 16:52:11.863623 18312 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0828 16:52:11.872713 18312 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0828 16:52:11.953265 18312 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0828 16:52:11.965312 18312 certs.go:68] Setting up /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161 for IP: 192.168.49.2
I0828 16:52:11.965343 18312 certs.go:194] generating shared ca certs ...
I0828 16:52:11.965361 18312 certs.go:226] acquiring lock for ca certs: {Name:mka9c7d226387ff2d3b302ddafee7ce1e10c1a14 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:11.965493 18312 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19529-10133/.minikube/ca.key
I0828 16:52:12.338385 18312 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19529-10133/.minikube/ca.crt ...
I0828 16:52:12.338417 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/ca.crt: {Name:mkdb555fe22045b941c1e1efb5c95fcc90c2252e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:12.338609 18312 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19529-10133/.minikube/ca.key ...
I0828 16:52:12.338625 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/ca.key: {Name:mk5e4eda7f00cfb5dff06377ac05354b4cc6c2f9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:12.338727 18312 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19529-10133/.minikube/proxy-client-ca.key
I0828 16:52:12.533749 18312 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19529-10133/.minikube/proxy-client-ca.crt ...
I0828 16:52:12.533777 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/proxy-client-ca.crt: {Name:mk1006c962adcfd73298ea544cb158eda977004a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:12.533951 18312 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19529-10133/.minikube/proxy-client-ca.key ...
I0828 16:52:12.533965 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/proxy-client-ca.key: {Name:mkc57bb2a9b62c62b06e43c4fe8c998e75d2dfb2 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:12.534053 18312 certs.go:256] generating profile certs ...
I0828 16:52:12.534120 18312 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/client.key
I0828 16:52:12.534139 18312 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/client.crt with IP's: []
I0828 16:52:12.753385 18312 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/client.crt ...
I0828 16:52:12.753411 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/client.crt: {Name:mkde35018ab4b9ed92b71ad70ba7e19f7c67a4e9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:12.753580 18312 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/client.key ...
I0828 16:52:12.753595 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/client.key: {Name:mk4d5c45d7082d56f5ffb7c38a363137607cbf9c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:12.753686 18312 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.key.db36f37b
I0828 16:52:12.753709 18312 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.crt.db36f37b with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0828 16:52:13.163800 18312 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.crt.db36f37b ...
I0828 16:52:13.163833 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.crt.db36f37b: {Name:mk15a42891e68409d682debfbef0fe8e4e1676d4 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:13.164009 18312 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.key.db36f37b ...
I0828 16:52:13.164027 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.key.db36f37b: {Name:mk6e5867aea31ca110b7a03bfd4956fb96cbd2a5 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:13.164122 18312 certs.go:381] copying /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.crt.db36f37b -> /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.crt
I0828 16:52:13.164249 18312 certs.go:385] copying /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.key.db36f37b -> /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.key
I0828 16:52:13.164325 18312 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/proxy-client.key
I0828 16:52:13.164355 18312 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/proxy-client.crt with IP's: []
I0828 16:52:13.333851 18312 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/proxy-client.crt ...
I0828 16:52:13.333887 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/proxy-client.crt: {Name:mk948f91f9e7e3d21b25d8f895ddb7b75eb8d6da Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:13.334070 18312 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/proxy-client.key ...
I0828 16:52:13.334085 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/proxy-client.key: {Name:mk7ddc9d103a54ba8d21e7eea47cc53f38fa1a6b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:13.334271 18312 certs.go:484] found cert: /home/jenkins/minikube-integration/19529-10133/.minikube/certs/ca-key.pem (1679 bytes)
I0828 16:52:13.334315 18312 certs.go:484] found cert: /home/jenkins/minikube-integration/19529-10133/.minikube/certs/ca.pem (1082 bytes)
I0828 16:52:13.334356 18312 certs.go:484] found cert: /home/jenkins/minikube-integration/19529-10133/.minikube/certs/cert.pem (1123 bytes)
I0828 16:52:13.334391 18312 certs.go:484] found cert: /home/jenkins/minikube-integration/19529-10133/.minikube/certs/key.pem (1679 bytes)
I0828 16:52:13.334937 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0828 16:52:13.355847 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0828 16:52:13.376039 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0828 16:52:13.396142 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0828 16:52:13.416218 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0828 16:52:13.436370 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0828 16:52:13.456867 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0828 16:52:13.477295 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/profiles/addons-001161/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I0828 16:52:13.499391 18312 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19529-10133/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0828 16:52:13.521419 18312 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0828 16:52:13.536413 18312 ssh_runner.go:195] Run: openssl version
I0828 16:52:13.541224 18312 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0828 16:52:13.549555 18312 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0828 16:52:13.552567 18312 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Aug 28 16:52 /usr/share/ca-certificates/minikubeCA.pem
I0828 16:52:13.552625 18312 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0828 16:52:13.558722 18312 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0828 16:52:13.566698 18312 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0828 16:52:13.569642 18312 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0828 16:52:13.569700 18312 kubeadm.go:392] StartCluster: {Name:addons-001161 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724775115-19521@sha256:5e61ebc6e68d69e31cadead040aa9b41aa36d281b29a7d562fa41077c99ed3ce Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-001161 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0828 16:52:13.569835 18312 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0828 16:52:13.585518 18312 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0828 16:52:13.593030 18312 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0828 16:52:13.600464 18312 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0828 16:52:13.600513 18312 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0828 16:52:13.607465 18312 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0828 16:52:13.607481 18312 kubeadm.go:157] found existing configuration files:
I0828 16:52:13.607517 18312 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0828 16:52:13.614483 18312 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0828 16:52:13.614533 18312 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0828 16:52:13.621330 18312 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0828 16:52:13.628389 18312 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0828 16:52:13.628446 18312 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0828 16:52:13.635303 18312 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0828 16:52:13.642459 18312 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0828 16:52:13.642507 18312 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0828 16:52:13.649524 18312 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0828 16:52:13.656799 18312 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0828 16:52:13.656856 18312 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0828 16:52:13.664003 18312 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.0:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0828 16:52:13.697406 18312 kubeadm.go:310] W0828 16:52:13.696697 1933 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0828 16:52:13.697872 18312 kubeadm.go:310] W0828 16:52:13.697315 1933 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0828 16:52:13.718337 18312 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1067-gcp\n", err: exit status 1
I0828 16:52:13.765516 18312 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0828 16:52:22.523176 18312 kubeadm.go:310] [init] Using Kubernetes version: v1.31.0
I0828 16:52:22.523252 18312 kubeadm.go:310] [preflight] Running pre-flight checks
I0828 16:52:22.523337 18312 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0828 16:52:22.523428 18312 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1067-gcp[0m
I0828 16:52:22.523491 18312 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0828 16:52:22.523556 18312 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0828 16:52:22.523638 18312 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0828 16:52:22.523712 18312 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0828 16:52:22.523770 18312 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0828 16:52:22.523844 18312 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0828 16:52:22.523902 18312 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0828 16:52:22.523975 18312 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0828 16:52:22.524034 18312 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0828 16:52:22.524103 18312 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0828 16:52:22.524197 18312 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0828 16:52:22.524353 18312 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0828 16:52:22.524463 18312 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0828 16:52:22.524563 18312 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0828 16:52:22.526252 18312 out.go:235] - Generating certificates and keys ...
I0828 16:52:22.526379 18312 kubeadm.go:310] [certs] Using existing ca certificate authority
I0828 16:52:22.526456 18312 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0828 16:52:22.526538 18312 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0828 16:52:22.526635 18312 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0828 16:52:22.526736 18312 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0828 16:52:22.526834 18312 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0828 16:52:22.526903 18312 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0828 16:52:22.527044 18312 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-001161 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0828 16:52:22.527109 18312 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0828 16:52:22.527241 18312 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-001161 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0828 16:52:22.527348 18312 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0828 16:52:22.527435 18312 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0828 16:52:22.527498 18312 kubeadm.go:310] [certs] Generating "sa" key and public key
I0828 16:52:22.527681 18312 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0828 16:52:22.527762 18312 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0828 16:52:22.527842 18312 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0828 16:52:22.527903 18312 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0828 16:52:22.527958 18312 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0828 16:52:22.528004 18312 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0828 16:52:22.528095 18312 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0828 16:52:22.528176 18312 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0828 16:52:22.529568 18312 out.go:235] - Booting up control plane ...
I0828 16:52:22.529658 18312 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0828 16:52:22.529734 18312 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0828 16:52:22.529801 18312 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0828 16:52:22.529900 18312 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0828 16:52:22.529992 18312 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0828 16:52:22.530025 18312 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0828 16:52:22.530138 18312 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0828 16:52:22.530225 18312 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0828 16:52:22.530285 18312 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.00217618s
I0828 16:52:22.530443 18312 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0828 16:52:22.530506 18312 kubeadm.go:310] [api-check] The API server is healthy after 4.001406502s
I0828 16:52:22.530598 18312 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0828 16:52:22.530713 18312 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0828 16:52:22.530789 18312 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0828 16:52:22.530956 18312 kubeadm.go:310] [mark-control-plane] Marking the node addons-001161 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0828 16:52:22.531009 18312 kubeadm.go:310] [bootstrap-token] Using token: h40zq7.6uxoj7h43nr711vo
I0828 16:52:22.532488 18312 out.go:235] - Configuring RBAC rules ...
I0828 16:52:22.532587 18312 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0828 16:52:22.532674 18312 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0828 16:52:22.532788 18312 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0828 16:52:22.532899 18312 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0828 16:52:22.533005 18312 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0828 16:52:22.533082 18312 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0828 16:52:22.533169 18312 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0828 16:52:22.533204 18312 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0828 16:52:22.533244 18312 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0828 16:52:22.533252 18312 kubeadm.go:310]
I0828 16:52:22.533301 18312 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0828 16:52:22.533307 18312 kubeadm.go:310]
I0828 16:52:22.533389 18312 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0828 16:52:22.533397 18312 kubeadm.go:310]
I0828 16:52:22.533422 18312 kubeadm.go:310] mkdir -p $HOME/.kube
I0828 16:52:22.533475 18312 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0828 16:52:22.533521 18312 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0828 16:52:22.533528 18312 kubeadm.go:310]
I0828 16:52:22.533568 18312 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0828 16:52:22.533574 18312 kubeadm.go:310]
I0828 16:52:22.533654 18312 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0828 16:52:22.533664 18312 kubeadm.go:310]
I0828 16:52:22.533723 18312 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0828 16:52:22.533824 18312 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0828 16:52:22.533925 18312 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0828 16:52:22.533933 18312 kubeadm.go:310]
I0828 16:52:22.534003 18312 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0828 16:52:22.534091 18312 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0828 16:52:22.534100 18312 kubeadm.go:310]
I0828 16:52:22.534166 18312 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token h40zq7.6uxoj7h43nr711vo \
I0828 16:52:22.534249 18312 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:a2c65f3f04f1b955be29fc7b8d8d349cd2dde3f9560501fc7a45c72f07d63b5b \
I0828 16:52:22.534268 18312 kubeadm.go:310] --control-plane
I0828 16:52:22.534274 18312 kubeadm.go:310]
I0828 16:52:22.534357 18312 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0828 16:52:22.534367 18312 kubeadm.go:310]
I0828 16:52:22.534464 18312 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token h40zq7.6uxoj7h43nr711vo \
I0828 16:52:22.534581 18312 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:a2c65f3f04f1b955be29fc7b8d8d349cd2dde3f9560501fc7a45c72f07d63b5b
I0828 16:52:22.534603 18312 cni.go:84] Creating CNI manager for ""
I0828 16:52:22.534619 18312 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0828 16:52:22.536347 18312 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0828 16:52:22.537552 18312 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0828 16:52:22.546261 18312 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0828 16:52:22.562752 18312 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0828 16:52:22.562822 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:22.562828 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-001161 minikube.k8s.io/updated_at=2024_08_28T16_52_22_0700 minikube.k8s.io/version=v1.33.1 minikube.k8s.io/commit=6f256f0bf490fd67de29a75a245d072e85b1b216 minikube.k8s.io/name=addons-001161 minikube.k8s.io/primary=true
I0828 16:52:22.643985 18312 ops.go:34] apiserver oom_adj: -16
I0828 16:52:22.644110 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:23.144365 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:23.644558 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:24.144992 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:24.644359 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:25.144353 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:25.644133 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:26.144606 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:26.644344 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:27.144147 18312 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0828 16:52:27.203645 18312 kubeadm.go:1113] duration metric: took 4.640882853s to wait for elevateKubeSystemPrivileges
I0828 16:52:27.203681 18312 kubeadm.go:394] duration metric: took 13.63398425s to StartCluster
I0828 16:52:27.203700 18312 settings.go:142] acquiring lock: {Name:mk219caa882a833007e0a46e6cc7216d561322be Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:27.203801 18312 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19529-10133/kubeconfig
I0828 16:52:27.204181 18312 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19529-10133/kubeconfig: {Name:mke85e0e520c0d04b61dde0ff6896e2be162f4cb Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0828 16:52:27.204399 18312 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0828 16:52:27.204418 18312 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0828 16:52:27.204477 18312 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0828 16:52:27.204594 18312 config.go:182] Loaded profile config "addons-001161": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0828 16:52:27.204605 18312 addons.go:69] Setting ingress-dns=true in profile "addons-001161"
I0828 16:52:27.204617 18312 addons.go:69] Setting storage-provisioner=true in profile "addons-001161"
I0828 16:52:27.204639 18312 addons.go:234] Setting addon ingress-dns=true in "addons-001161"
I0828 16:52:27.204644 18312 addons.go:234] Setting addon storage-provisioner=true in "addons-001161"
I0828 16:52:27.204635 18312 addons.go:69] Setting registry=true in profile "addons-001161"
I0828 16:52:27.204652 18312 addons.go:69] Setting cloud-spanner=true in profile "addons-001161"
I0828 16:52:27.204653 18312 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-001161"
I0828 16:52:27.204671 18312 addons.go:234] Setting addon cloud-spanner=true in "addons-001161"
I0828 16:52:27.204683 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.204684 18312 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-001161"
I0828 16:52:27.204687 18312 addons.go:69] Setting gcp-auth=true in profile "addons-001161"
I0828 16:52:27.204692 18312 addons.go:69] Setting ingress=true in profile "addons-001161"
I0828 16:52:27.204695 18312 addons.go:69] Setting metrics-server=true in profile "addons-001161"
I0828 16:52:27.204705 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.204712 18312 addons.go:234] Setting addon ingress=true in "addons-001161"
I0828 16:52:27.204715 18312 addons.go:234] Setting addon metrics-server=true in "addons-001161"
I0828 16:52:27.204717 18312 addons.go:69] Setting volumesnapshots=true in profile "addons-001161"
I0828 16:52:27.204731 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.204737 18312 addons.go:234] Setting addon volumesnapshots=true in "addons-001161"
I0828 16:52:27.204749 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.204769 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.205018 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.205139 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.205177 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.205192 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.204635 18312 addons.go:69] Setting default-storageclass=true in profile "addons-001161"
I0828 16:52:27.205195 18312 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-001161"
I0828 16:52:27.205219 18312 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-001161"
I0828 16:52:27.205240 18312 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-001161"
I0828 16:52:27.205264 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.205419 18312 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-001161"
I0828 16:52:27.205449 18312 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-001161"
I0828 16:52:27.205451 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.205473 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.205193 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.205682 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.205911 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.205187 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.204683 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.208476 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.204672 18312 addons.go:234] Setting addon registry=true in "addons-001161"
I0828 16:52:27.210415 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.210898 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.204667 18312 addons.go:69] Setting helm-tiller=true in profile "addons-001161"
I0828 16:52:27.211101 18312 addons.go:234] Setting addon helm-tiller=true in "addons-001161"
I0828 16:52:27.211176 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.204707 18312 mustload.go:65] Loading cluster: addons-001161
I0828 16:52:27.212496 18312 out.go:177] * Verifying Kubernetes components...
I0828 16:52:27.212690 18312 config.go:182] Loaded profile config "addons-001161": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0828 16:52:27.204644 18312 addons.go:69] Setting volcano=true in profile "addons-001161"
I0828 16:52:27.212888 18312 addons.go:234] Setting addon volcano=true in "addons-001161"
I0828 16:52:27.212927 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.213083 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.213439 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.204603 18312 addons.go:69] Setting yakd=true in profile "addons-001161"
I0828 16:52:27.214274 18312 addons.go:234] Setting addon yakd=true in "addons-001161"
I0828 16:52:27.214332 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.214797 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.215043 18312 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0828 16:52:27.204684 18312 addons.go:69] Setting inspektor-gadget=true in profile "addons-001161"
I0828 16:52:27.215771 18312 addons.go:234] Setting addon inspektor-gadget=true in "addons-001161"
I0828 16:52:27.215811 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.216251 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.221146 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.240310 18312 addons.go:234] Setting addon default-storageclass=true in "addons-001161"
I0828 16:52:27.240346 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.240937 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.245304 18312 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0828 16:52:27.247003 18312 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0828 16:52:27.247026 18312 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0828 16:52:27.247090 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.252321 18312 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0828 16:52:27.254081 18312 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0828 16:52:27.255281 18312 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0828 16:52:27.256636 18312 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0828 16:52:27.257924 18312 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0828 16:52:27.258050 18312 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0828 16:52:27.259325 18312 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0828 16:52:27.259371 18312 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0828 16:52:27.259388 18312 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0828 16:52:27.259404 18312 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0828 16:52:27.259623 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.260720 18312 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0828 16:52:27.260738 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0828 16:52:27.260784 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.262051 18312 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0828 16:52:27.263194 18312 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0828 16:52:27.264099 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.264399 18312 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0828 16:52:27.264414 18312 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0828 16:52:27.264455 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.271782 18312 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0828 16:52:27.273016 18312 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0828 16:52:27.274340 18312 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0828 16:52:27.275851 18312 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0828 16:52:27.275869 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0828 16:52:27.275918 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.277612 18312 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-001161"
I0828 16:52:27.277700 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:27.277837 18312 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0828 16:52:27.278335 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:27.279003 18312 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0828 16:52:27.279024 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0828 16:52:27.279069 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.283417 18312 out.go:177] - Using image docker.io/registry:2.8.3
I0828 16:52:27.284862 18312 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0828 16:52:27.286036 18312 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0828 16:52:27.286057 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0828 16:52:27.286107 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.286592 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.289867 18312 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0828 16:52:27.290694 18312 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0828 16:52:27.291562 18312 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0828 16:52:27.291582 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0828 16:52:27.291618 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.292273 18312 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0828 16:52:27.292287 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0828 16:52:27.292328 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.294476 18312 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0828 16:52:27.294491 18312 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0828 16:52:27.294539 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.303173 18312 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.31.0
I0828 16:52:27.307481 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.308574 18312 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0828 16:52:27.308603 18312 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0828 16:52:27.308667 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.311371 18312 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0828 16:52:27.311516 18312 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0828 16:52:27.312940 18312 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0828 16:52:27.312961 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0828 16:52:27.313020 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.313266 18312 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0828 16:52:27.320420 18312 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0828 16:52:27.323015 18312 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0828 16:52:27.323043 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0828 16:52:27.323108 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.345951 18312 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0828 16:52:27.347175 18312 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0828 16:52:27.347195 18312 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0828 16:52:27.347257 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.348214 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.355346 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.363173 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.370765 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.378694 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.378936 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.379294 18312 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0828 16:52:27.382451 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.384082 18312 out.go:177] - Using image docker.io/busybox:stable
I0828 16:52:27.387140 18312 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0828 16:52:27.387157 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0828 16:52:27.387208 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:27.392200 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.393398 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.400766 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.404212 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.405552 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:27.408350 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
W0828 16:52:27.432480 18312 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0828 16:52:27.432510 18312 retry.go:31] will retry after 209.955773ms: ssh: handshake failed: EOF
W0828 16:52:27.432591 18312 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0828 16:52:27.432600 18312 retry.go:31] will retry after 342.756645ms: ssh: handshake failed: EOF
I0828 16:52:27.730681 18312 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0828 16:52:27.730822 18312 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0828 16:52:27.745763 18312 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0828 16:52:27.745792 18312 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0828 16:52:27.752668 18312 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0828 16:52:27.752694 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0828 16:52:27.837576 18312 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0828 16:52:27.837603 18312 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0828 16:52:27.840435 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0828 16:52:27.939782 18312 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0828 16:52:27.939883 18312 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0828 16:52:27.950395 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0828 16:52:27.952158 18312 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0828 16:52:27.952182 18312 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0828 16:52:28.035040 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0828 16:52:28.046978 18312 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0828 16:52:28.047006 18312 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0828 16:52:28.141781 18312 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0828 16:52:28.141889 18312 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0828 16:52:28.146301 18312 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0828 16:52:28.146327 18312 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0828 16:52:28.150626 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0828 16:52:28.232079 18312 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0828 16:52:28.232113 18312 ssh_runner.go:362] scp helm-tiller/helm-tiller-rbac.yaml --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0828 16:52:28.330224 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0828 16:52:28.348632 18312 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0828 16:52:28.348720 18312 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0828 16:52:28.434266 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0828 16:52:28.441007 18312 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0828 16:52:28.441089 18312 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0828 16:52:28.451105 18312 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0828 16:52:28.451139 18312 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0828 16:52:28.451377 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0828 16:52:28.452453 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0828 16:52:28.453297 18312 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0828 16:52:28.455046 18312 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0828 16:52:28.529822 18312 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0828 16:52:28.529908 18312 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0828 16:52:28.637313 18312 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0828 16:52:28.637400 18312 ssh_runner.go:362] scp helm-tiller/helm-tiller-svc.yaml --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0828 16:52:28.739184 18312 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0828 16:52:28.739266 18312 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0828 16:52:28.751052 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0828 16:52:28.835700 18312 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0828 16:52:28.835732 18312 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0828 16:52:28.836947 18312 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0828 16:52:28.836971 18312 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0828 16:52:29.032031 18312 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0828 16:52:29.032059 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0828 16:52:29.046481 18312 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0828 16:52:29.046562 18312 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0828 16:52:29.129980 18312 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0828 16:52:29.130066 18312 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0828 16:52:29.151555 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0828 16:52:29.230895 18312 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0828 16:52:29.230971 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0828 16:52:29.330934 18312 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0828 16:52:29.331013 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0828 16:52:29.337521 18312 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0828 16:52:29.337600 18312 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0828 16:52:29.444521 18312 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.713670691s)
I0828 16:52:29.444959 18312 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.714236948s)
I0828 16:52:29.445091 18312 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0828 16:52:29.446998 18312 node_ready.go:35] waiting up to 6m0s for node "addons-001161" to be "Ready" ...
I0828 16:52:29.450540 18312 node_ready.go:49] node "addons-001161" has status "Ready":"True"
I0828 16:52:29.450603 18312 node_ready.go:38] duration metric: took 3.544213ms for node "addons-001161" to be "Ready" ...
I0828 16:52:29.450625 18312 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0828 16:52:29.531469 18312 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0828 16:52:29.531497 18312 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0828 16:52:29.540902 18312 pod_ready.go:79] waiting up to 6m0s for pod "coredns-6f6b679f8f-62gds" in "kube-system" namespace to be "Ready" ...
I0828 16:52:29.630984 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0828 16:52:29.734232 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0828 16:52:29.746946 18312 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0828 16:52:29.747033 18312 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0828 16:52:29.844224 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0828 16:52:29.950392 18312 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-001161" context rescaled to 1 replicas
I0828 16:52:30.029115 18312 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0828 16:52:30.029207 18312 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0828 16:52:30.131260 18312 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0828 16:52:30.131349 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0828 16:52:30.549867 18312 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0828 16:52:30.549902 18312 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0828 16:52:30.745649 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (2.905176737s)
I0828 16:52:30.941170 18312 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0828 16:52:30.941267 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0828 16:52:31.436521 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0828 16:52:31.540414 18312 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0828 16:52:31.540506 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0828 16:52:31.634285 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-62gds" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:31.933445 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (3.98300632s)
I0828 16:52:31.933610 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (3.898520166s)
I0828 16:52:32.242907 18312 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0828 16:52:32.242994 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0828 16:52:32.836673 18312 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0828 16:52:32.836762 18312 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0828 16:52:32.929264 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (4.778540551s)
I0828 16:52:32.929651 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (4.599395011s)
W0828 16:52:32.952095 18312 out.go:270] ! Enabling 'storage-provisioner-rancher' returned an error: running callbacks: [Error making local-path the default storage class: Error while marking storage class local-path as default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
I0828 16:52:33.435787 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0828 16:52:33.645250 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-62gds" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:34.334839 18312 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0828 16:52:34.335005 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:34.361155 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:35.445724 18312 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0828 16:52:35.749771 18312 addons.go:234] Setting addon gcp-auth=true in "addons-001161"
I0828 16:52:35.749901 18312 host.go:66] Checking if "addons-001161" exists ...
I0828 16:52:35.750459 18312 cli_runner.go:164] Run: docker container inspect addons-001161 --format={{.State.Status}}
I0828 16:52:35.767857 18312 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0828 16:52:35.767904 18312 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-001161
I0828 16:52:35.784423 18312 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19529-10133/.minikube/machines/addons-001161/id_rsa Username:docker}
I0828 16:52:36.048206 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-62gds" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:38.050008 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-62gds" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:39.735559 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.301178028s)
I0828 16:52:39.735653 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (11.284250044s)
I0828 16:52:39.735693 18312 addons.go:475] Verifying addon ingress=true in "addons-001161"
I0828 16:52:39.735871 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (11.283390462s)
I0828 16:52:39.735956 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (10.984874302s)
I0828 16:52:39.735981 18312 addons.go:475] Verifying addon metrics-server=true in "addons-001161"
I0828 16:52:39.736026 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (10.584445707s)
I0828 16:52:39.736130 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (10.105108539s)
W0828 16:52:39.736154 18312 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0828 16:52:39.736171 18312 retry.go:31] will retry after 256.361902ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0828 16:52:39.736254 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (10.001987284s)
I0828 16:52:39.736368 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (9.892053802s)
I0828 16:52:39.737022 18312 addons.go:475] Verifying addon registry=true in "addons-001161"
I0828 16:52:39.736454 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (8.299842831s)
I0828 16:52:39.737397 18312 out.go:177] * Verifying ingress addon...
I0828 16:52:39.739153 18312 out.go:177] * Verifying registry addon...
I0828 16:52:39.739164 18312 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-001161 service yakd-dashboard -n yakd-dashboard
I0828 16:52:39.740176 18312 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0828 16:52:39.741647 18312 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0828 16:52:39.744565 18312 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0828 16:52:39.744588 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:39.844179 18312 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0828 16:52:39.844217 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:39.993615 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0828 16:52:40.136637 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-62gds" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:40.247713 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:40.248073 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:40.746334 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:40.747340 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:40.843299 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (7.407432738s)
I0828 16:52:40.843343 18312 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-001161"
I0828 16:52:40.843398 18312 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.075442393s)
I0828 16:52:40.844894 18312 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0828 16:52:40.844898 18312 out.go:177] * Verifying csi-hostpath-driver addon...
I0828 16:52:40.847761 18312 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0828 16:52:40.849713 18312 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0828 16:52:40.851710 18312 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0828 16:52:40.851733 18312 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0828 16:52:40.856573 18312 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0828 16:52:40.856634 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:40.941341 18312 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0828 16:52:40.941406 18312 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0828 16:52:41.029153 18312 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0828 16:52:41.029179 18312 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0828 16:52:41.050660 18312 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0828 16:52:41.245062 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:41.245165 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:41.354627 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:41.745566 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:41.746420 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:41.853462 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:42.048350 18312 pod_ready.go:98] pod "coredns-6f6b679f8f-62gds" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-08-28 16:52:41 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-08-28 16:52:27 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-08-28 16:52:27 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-08-28 16:52:27 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-08-28 16:52:27 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.49.2 HostIPs:[{IP:192.168.49.2
}] PodIP:10.244.0.2 PodIPs:[{IP:10.244.0.2}] StartTime:2024-08-28 16:52:27 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-08-28 16:52:30 +0000 UTC,FinishedAt:2024-08-28 16:52:40 +0000 UTC,ContainerID:docker://04dae1b69b9b473113c43b9ab3baef36e3b844ef911fdb46f06d36bb9045bb90,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.1 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1 ContainerID:docker://04dae1b69b9b473113c43b9ab3baef36e3b844ef911fdb46f06d36bb9045bb90 Started:0xc00239f610 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0xc0023b0880} {Name:kube-api-access-fvml2 MountPath:/var/run/secrets/kubernetes.io/serviceaccount
ReadOnly:true RecursiveReadOnly:0xc0023b0890}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
I0828 16:52:42.048379 18312 pod_ready.go:82] duration metric: took 12.507444905s for pod "coredns-6f6b679f8f-62gds" in "kube-system" namespace to be "Ready" ...
E0828 16:52:42.048390 18312 pod_ready.go:67] WaitExtra: waitPodCondition: pod "coredns-6f6b679f8f-62gds" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-08-28 16:52:41 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-08-28 16:52:27 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-08-28 16:52:27 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-08-28 16:52:27 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-08-28 16:52:27 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.4
9.2 HostIPs:[{IP:192.168.49.2}] PodIP:10.244.0.2 PodIPs:[{IP:10.244.0.2}] StartTime:2024-08-28 16:52:27 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-08-28 16:52:30 +0000 UTC,FinishedAt:2024-08-28 16:52:40 +0000 UTC,ContainerID:docker://04dae1b69b9b473113c43b9ab3baef36e3b844ef911fdb46f06d36bb9045bb90,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.1 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1 ContainerID:docker://04dae1b69b9b473113c43b9ab3baef36e3b844ef911fdb46f06d36bb9045bb90 Started:0xc00239f610 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0xc0023b0880} {Name:kube-api-access-fvml2 MountPath:/var/run/secrets
/kubernetes.io/serviceaccount ReadOnly:true RecursiveReadOnly:0xc0023b0890}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
I0828 16:52:42.048400 18312 pod_ready.go:79] waiting up to 6m0s for pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace to be "Ready" ...
I0828 16:52:42.245631 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:42.246115 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:42.352557 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:42.551203 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.557540006s)
I0828 16:52:42.658699 18312 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.607938985s)
I0828 16:52:42.660454 18312 addons.go:475] Verifying addon gcp-auth=true in "addons-001161"
I0828 16:52:42.663665 18312 out.go:177] * Verifying gcp-auth addon...
I0828 16:52:42.665748 18312 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0828 16:52:42.667601 18312 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0828 16:52:42.769106 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:42.769678 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:42.870344 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:43.244829 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:43.245349 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:43.352273 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:43.744080 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:43.745485 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:43.852049 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:44.054780 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:44.243994 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:44.245222 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:44.352928 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:44.745139 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:44.745759 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:44.851558 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:45.244318 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:45.244564 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:45.352451 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:45.770139 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:45.770542 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:45.852741 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:46.267830 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:46.268154 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:46.351873 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:46.553926 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:46.744635 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:46.744957 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:46.851965 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:47.244283 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:47.244396 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:47.351878 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:47.744339 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:47.744487 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:47.852032 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:48.244178 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:48.244583 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:48.352048 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:48.744089 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:48.744299 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:48.851695 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:49.054058 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:49.244984 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:49.244992 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:49.352594 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:49.770684 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:49.770938 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:49.852545 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:50.244702 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:50.244902 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:50.352001 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:50.744450 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:50.744627 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:50.852259 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:51.244124 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:51.244522 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:51.351824 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:51.554087 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:51.744845 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:51.745025 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:51.852525 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:52.245215 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:52.245251 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:52.353083 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:52.744437 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:52.745232 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:52.851877 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:53.244593 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:53.244639 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:53.351738 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:53.743977 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:53.744283 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:53.851238 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:54.053323 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:54.244275 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:54.244275 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:54.351253 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:54.744189 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:54.745185 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:54.852118 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:55.244215 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:55.244735 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:55.351961 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:55.743314 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:55.744688 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:55.851838 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:56.053716 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:56.243577 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:56.244978 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:56.352014 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:56.744904 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:56.744944 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:56.851796 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:57.243434 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:57.244832 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:57.354067 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:57.770552 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:57.771064 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:57.852108 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:58.054186 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:52:58.244686 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:58.245206 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:58.352186 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:58.744469 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:58.744721 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:58.852100 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:59.270295 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:59.270438 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:59.352567 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:52:59.743449 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:52:59.744683 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:52:59.851973 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:00.244423 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:00.244777 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:00.352429 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:00.554513 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:53:00.770149 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:00.770518 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:00.871319 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:01.244040 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:01.244183 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:01.351336 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:01.743492 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:01.745044 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:01.852449 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:02.244140 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:02.244269 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:02.352519 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:02.770809 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:02.771165 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:02.852257 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:03.054481 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:53:03.244569 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:03.244654 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:03.353201 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:03.826631 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:03.826955 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:03.929836 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:04.243901 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:04.244979 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:04.352164 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:04.744220 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:04.746979 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:04.852286 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:05.054597 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:53:05.244967 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:05.245479 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:05.351832 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:05.744092 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:05.744534 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:05.852712 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:06.244160 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:06.244433 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:06.351889 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:06.744082 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:06.745651 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:06.851395 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:07.244490 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:07.244592 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:07.351495 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:07.553555 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:53:07.744434 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:07.744776 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:07.852116 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:08.244775 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:08.244875 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:08.352640 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:08.744540 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:08.744935 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:08.980064 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:09.270910 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:09.271159 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:09.351894 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:09.553841 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:53:09.744179 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:09.744794 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:09.851832 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:10.245131 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:10.245370 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:10.351682 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:10.744283 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:10.744646 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:10.852434 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:11.244109 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:11.244310 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:11.352894 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:11.554113 18312 pod_ready.go:103] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"False"
I0828 16:53:11.744576 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:11.745624 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:11.852348 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:12.054605 18312 pod_ready.go:93] pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace has status "Ready":"True"
I0828 16:53:12.054634 18312 pod_ready.go:82] duration metric: took 30.006221459s for pod "coredns-6f6b679f8f-ls4v7" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.054652 18312 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-001161" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.058631 18312 pod_ready.go:93] pod "etcd-addons-001161" in "kube-system" namespace has status "Ready":"True"
I0828 16:53:12.058688 18312 pod_ready.go:82] duration metric: took 4.025877ms for pod "etcd-addons-001161" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.058705 18312 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-001161" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.064489 18312 pod_ready.go:93] pod "kube-apiserver-addons-001161" in "kube-system" namespace has status "Ready":"True"
I0828 16:53:12.064510 18312 pod_ready.go:82] duration metric: took 5.79578ms for pod "kube-apiserver-addons-001161" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.064519 18312 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-001161" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.068448 18312 pod_ready.go:93] pod "kube-controller-manager-addons-001161" in "kube-system" namespace has status "Ready":"True"
I0828 16:53:12.068466 18312 pod_ready.go:82] duration metric: took 3.940952ms for pod "kube-controller-manager-addons-001161" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.068475 18312 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-672sn" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.072328 18312 pod_ready.go:93] pod "kube-proxy-672sn" in "kube-system" namespace has status "Ready":"True"
I0828 16:53:12.072347 18312 pod_ready.go:82] duration metric: took 3.865865ms for pod "kube-proxy-672sn" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.072355 18312 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-001161" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.244935 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:12.245194 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:12.352483 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:12.452896 18312 pod_ready.go:93] pod "kube-scheduler-addons-001161" in "kube-system" namespace has status "Ready":"True"
I0828 16:53:12.452922 18312 pod_ready.go:82] duration metric: took 380.558983ms for pod "kube-scheduler-addons-001161" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.452936 18312 pod_ready.go:79] waiting up to 6m0s for pod "nvidia-device-plugin-daemonset-w74pg" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.743911 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:12.744381 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:12.852278 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:12.852655 18312 pod_ready.go:93] pod "nvidia-device-plugin-daemonset-w74pg" in "kube-system" namespace has status "Ready":"True"
I0828 16:53:12.852674 18312 pod_ready.go:82] duration metric: took 399.731105ms for pod "nvidia-device-plugin-daemonset-w74pg" in "kube-system" namespace to be "Ready" ...
I0828 16:53:12.852682 18312 pod_ready.go:39] duration metric: took 43.402036712s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0828 16:53:12.852702 18312 api_server.go:52] waiting for apiserver process to appear ...
I0828 16:53:12.852748 18312 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0828 16:53:12.866048 18312 api_server.go:72] duration metric: took 45.661594728s to wait for apiserver process to appear ...
I0828 16:53:12.866074 18312 api_server.go:88] waiting for apiserver healthz status ...
I0828 16:53:12.866091 18312 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0828 16:53:12.943704 18312 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0828 16:53:12.944726 18312 api_server.go:141] control plane version: v1.31.0
I0828 16:53:12.944747 18312 api_server.go:131] duration metric: took 78.666605ms to wait for apiserver health ...
I0828 16:53:12.944757 18312 system_pods.go:43] waiting for kube-system pods to appear ...
I0828 16:53:13.057336 18312 system_pods.go:59] 18 kube-system pods found
I0828 16:53:13.057368 18312 system_pods.go:61] "coredns-6f6b679f8f-ls4v7" [a285d539-6a5c-44ff-bef1-77c88b2d30f2] Running
I0828 16:53:13.057377 18312 system_pods.go:61] "csi-hostpath-attacher-0" [220215e7-73f2-40bc-91f2-a7609d6dc726] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0828 16:53:13.057383 18312 system_pods.go:61] "csi-hostpath-resizer-0" [bf6530d8-1bae-481f-b724-e8fa65fd1ee1] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0828 16:53:13.057390 18312 system_pods.go:61] "csi-hostpathplugin-7hp4n" [5ad27c5c-17a5-4633-82fc-2385b7bd4944] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0828 16:53:13.057396 18312 system_pods.go:61] "etcd-addons-001161" [bbd8c430-11b3-437d-b746-5f1303ebeaa6] Running
I0828 16:53:13.057401 18312 system_pods.go:61] "kube-apiserver-addons-001161" [a81ada16-fa22-4292-8c0e-e75556c25fd9] Running
I0828 16:53:13.057405 18312 system_pods.go:61] "kube-controller-manager-addons-001161" [dbcea47a-9d1f-405d-91ee-a37df5e51d0c] Running
I0828 16:53:13.057409 18312 system_pods.go:61] "kube-ingress-dns-minikube" [2e4ef842-1957-4b72-b3f1-d8be8e03d4d1] Running
I0828 16:53:13.057412 18312 system_pods.go:61] "kube-proxy-672sn" [a737d6c4-f8ea-43bc-b712-5f3aef4cedd4] Running
I0828 16:53:13.057415 18312 system_pods.go:61] "kube-scheduler-addons-001161" [7251a007-2ced-4b93-8864-938fb7c71309] Running
I0828 16:53:13.057419 18312 system_pods.go:61] "metrics-server-84c5f94fbc-mw2t4" [8dd70871-3102-43cf-8eaf-b438aeb1c66c] Running
I0828 16:53:13.057426 18312 system_pods.go:61] "nvidia-device-plugin-daemonset-w74pg" [0c9f48b4-c961-4444-978a-5ea6a57ae211] Running
I0828 16:53:13.057429 18312 system_pods.go:61] "registry-6fb4cdfc84-lbgl8" [243f93fd-9335-4ba8-a33c-d9ed0affd92f] Running
I0828 16:53:13.057436 18312 system_pods.go:61] "registry-proxy-vxpf8" [6d3d6e6b-090e-4033-9117-635faf64b5bf] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0828 16:53:13.057442 18312 system_pods.go:61] "snapshot-controller-56fcc65765-7v8qw" [4bfc7192-ef8b-4ccf-87f1-451168ab8fb6] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0828 16:53:13.057450 18312 system_pods.go:61] "snapshot-controller-56fcc65765-8pv8r" [794b01fa-aedf-4cc2-89b1-a6f1ad17e165] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0828 16:53:13.057454 18312 system_pods.go:61] "storage-provisioner" [40f7ba2f-8028-4db7-bce1-0795a5260a7c] Running
I0828 16:53:13.057457 18312 system_pods.go:61] "tiller-deploy-b48cc5f79-wqzh4" [80f8a125-9203-4e69-8c6c-ba6c41b51769] Running
I0828 16:53:13.057464 18312 system_pods.go:74] duration metric: took 112.701175ms to wait for pod list to return data ...
I0828 16:53:13.057473 18312 default_sa.go:34] waiting for default service account to be created ...
I0828 16:53:13.243542 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:13.245650 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:13.252342 18312 default_sa.go:45] found service account: "default"
I0828 16:53:13.252364 18312 default_sa.go:55] duration metric: took 194.882948ms for default service account to be created ...
I0828 16:53:13.252373 18312 system_pods.go:116] waiting for k8s-apps to be running ...
I0828 16:53:13.352336 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:13.457286 18312 system_pods.go:86] 18 kube-system pods found
I0828 16:53:13.457322 18312 system_pods.go:89] "coredns-6f6b679f8f-ls4v7" [a285d539-6a5c-44ff-bef1-77c88b2d30f2] Running
I0828 16:53:13.457337 18312 system_pods.go:89] "csi-hostpath-attacher-0" [220215e7-73f2-40bc-91f2-a7609d6dc726] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0828 16:53:13.457357 18312 system_pods.go:89] "csi-hostpath-resizer-0" [bf6530d8-1bae-481f-b724-e8fa65fd1ee1] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0828 16:53:13.457370 18312 system_pods.go:89] "csi-hostpathplugin-7hp4n" [5ad27c5c-17a5-4633-82fc-2385b7bd4944] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0828 16:53:13.457380 18312 system_pods.go:89] "etcd-addons-001161" [bbd8c430-11b3-437d-b746-5f1303ebeaa6] Running
I0828 16:53:13.457387 18312 system_pods.go:89] "kube-apiserver-addons-001161" [a81ada16-fa22-4292-8c0e-e75556c25fd9] Running
I0828 16:53:13.457396 18312 system_pods.go:89] "kube-controller-manager-addons-001161" [dbcea47a-9d1f-405d-91ee-a37df5e51d0c] Running
I0828 16:53:13.457402 18312 system_pods.go:89] "kube-ingress-dns-minikube" [2e4ef842-1957-4b72-b3f1-d8be8e03d4d1] Running
I0828 16:53:13.457412 18312 system_pods.go:89] "kube-proxy-672sn" [a737d6c4-f8ea-43bc-b712-5f3aef4cedd4] Running
I0828 16:53:13.457419 18312 system_pods.go:89] "kube-scheduler-addons-001161" [7251a007-2ced-4b93-8864-938fb7c71309] Running
I0828 16:53:13.457427 18312 system_pods.go:89] "metrics-server-84c5f94fbc-mw2t4" [8dd70871-3102-43cf-8eaf-b438aeb1c66c] Running
I0828 16:53:13.457433 18312 system_pods.go:89] "nvidia-device-plugin-daemonset-w74pg" [0c9f48b4-c961-4444-978a-5ea6a57ae211] Running
I0828 16:53:13.457442 18312 system_pods.go:89] "registry-6fb4cdfc84-lbgl8" [243f93fd-9335-4ba8-a33c-d9ed0affd92f] Running
I0828 16:53:13.457451 18312 system_pods.go:89] "registry-proxy-vxpf8" [6d3d6e6b-090e-4033-9117-635faf64b5bf] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0828 16:53:13.457467 18312 system_pods.go:89] "snapshot-controller-56fcc65765-7v8qw" [4bfc7192-ef8b-4ccf-87f1-451168ab8fb6] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0828 16:53:13.457476 18312 system_pods.go:89] "snapshot-controller-56fcc65765-8pv8r" [794b01fa-aedf-4cc2-89b1-a6f1ad17e165] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0828 16:53:13.457483 18312 system_pods.go:89] "storage-provisioner" [40f7ba2f-8028-4db7-bce1-0795a5260a7c] Running
I0828 16:53:13.457489 18312 system_pods.go:89] "tiller-deploy-b48cc5f79-wqzh4" [80f8a125-9203-4e69-8c6c-ba6c41b51769] Running
I0828 16:53:13.457502 18312 system_pods.go:126] duration metric: took 205.121543ms to wait for k8s-apps to be running ...
I0828 16:53:13.457514 18312 system_svc.go:44] waiting for kubelet service to be running ....
I0828 16:53:13.457563 18312 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0828 16:53:13.469367 18312 system_svc.go:56] duration metric: took 11.84487ms WaitForService to wait for kubelet
I0828 16:53:13.469400 18312 kubeadm.go:582] duration metric: took 46.264950494s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0828 16:53:13.469427 18312 node_conditions.go:102] verifying NodePressure condition ...
I0828 16:53:13.652107 18312 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0828 16:53:13.652131 18312 node_conditions.go:123] node cpu capacity is 8
I0828 16:53:13.652143 18312 node_conditions.go:105] duration metric: took 182.710565ms to run NodePressure ...
I0828 16:53:13.652153 18312 start.go:241] waiting for startup goroutines ...
I0828 16:53:13.745270 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:13.745640 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:13.854230 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:14.244216 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:14.244363 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:14.351437 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:14.744380 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:14.744664 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:14.851872 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:15.244634 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:15.244842 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:15.351791 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:15.744914 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:15.744963 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:15.852831 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:16.244766 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:16.245045 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:16.352604 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:16.744949 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:16.745194 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:16.852786 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:17.244128 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:17.245736 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:17.352357 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:17.744546 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:17.744805 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:17.852628 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:18.244638 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:18.244688 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:18.351507 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:18.744364 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0828 16:53:18.744703 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:18.852699 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:19.244916 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:19.246182 18312 kapi.go:107] duration metric: took 39.504532768s to wait for kubernetes.io/minikube-addons=registry ...
I0828 16:53:19.353241 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:19.770499 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:19.865413 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:20.244717 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:20.351754 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:20.832700 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:20.930628 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:21.243588 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:21.352154 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:21.744123 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:21.851709 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:22.243980 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:22.352193 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:22.744500 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:22.853018 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:23.245915 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:23.367629 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:23.744201 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:23.852528 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:24.244654 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:24.351692 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:24.743969 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:24.852197 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:25.243943 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:25.352455 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:25.743907 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:25.852282 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:26.244823 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:26.353111 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:26.743316 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:26.852444 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:27.244214 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:27.351963 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:27.743580 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:27.851140 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:28.244401 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:28.351926 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:28.744267 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:28.852455 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:29.244573 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:29.390485 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:29.744578 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:29.852626 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:30.271038 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:30.351369 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:30.744040 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:30.852406 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:31.244263 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:31.352594 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:31.743470 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:31.852885 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:32.244552 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:32.353204 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:32.744177 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:32.852384 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:33.244252 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:33.353092 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:33.771780 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:33.873255 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:34.244740 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:34.351745 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:34.770240 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:34.870874 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:35.245106 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:35.352184 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:35.743911 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:35.852111 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:36.244415 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:36.352564 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:36.743501 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:36.851255 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:37.245437 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:37.352527 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:37.744437 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:37.852128 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:38.243732 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:38.351268 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0828 16:53:38.743331 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:38.852671 18312 kapi.go:107] duration metric: took 58.004909072s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0828 16:53:39.243991 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:39.743729 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:40.244176 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:40.744376 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:41.244326 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:41.743383 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:42.244943 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:42.744518 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:43.244067 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:43.744563 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:44.244483 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:44.744834 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:45.244817 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:45.790039 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:46.270947 18312 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0828 16:53:46.770911 18312 kapi.go:107] duration metric: took 1m7.030733637s to wait for app.kubernetes.io/name=ingress-nginx ...
I0828 16:54:05.168984 18312 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0828 16:54:05.169004 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:05.668969 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:06.169571 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:06.669102 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:07.169022 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:07.669094 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:08.169275 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:08.669566 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:09.169300 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:09.668842 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:10.169081 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:10.668810 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:11.169420 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:11.669268 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:12.169045 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:12.669071 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:13.169324 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:13.669114 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:14.168382 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:14.669704 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:15.168462 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:15.669268 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:16.169706 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:16.668780 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:17.168526 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:17.669720 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:18.169322 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:18.669663 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:19.168972 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:19.668502 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:20.169845 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:20.668905 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:21.169061 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:21.668777 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:22.169004 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:22.669073 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:23.168875 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:23.668992 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:24.169137 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:24.668726 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:25.168788 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:25.668536 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:26.169192 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:26.669052 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:27.168821 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:27.668893 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:28.168492 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:28.669494 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:29.169678 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:29.668903 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:30.168930 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:30.669085 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:31.169228 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:31.669044 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:32.168756 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:32.669058 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:33.169121 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:33.669186 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:34.169380 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:34.669458 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:35.169267 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:35.669026 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:36.169086 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:36.668651 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:37.169252 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:37.669209 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:38.168928 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:38.669179 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:39.168662 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:39.669310 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:40.169217 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:40.669305 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:41.171443 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:41.669102 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:42.169041 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:42.668873 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:43.168451 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:43.669248 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:44.169102 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:44.669168 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:45.169383 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:45.669082 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:46.170179 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:46.669011 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:47.168611 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:47.668887 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:48.168601 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:48.669303 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:49.169150 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:49.669233 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:50.169386 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:50.669495 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:51.169768 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:51.669284 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:52.169111 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:52.669080 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:53.168961 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:53.669119 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:54.168907 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:54.668875 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:55.168881 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:55.668667 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:56.168894 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:56.668935 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:57.168651 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:57.668448 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:58.169366 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:58.669583 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:59.169477 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:54:59.669263 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:00.169486 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:00.669683 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:01.168816 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:01.669077 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:02.168789 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:02.668379 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:03.169319 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:03.669125 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:04.169262 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:04.669137 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:05.169021 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:05.668610 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:06.169293 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:06.669446 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:07.169239 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:07.668986 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:08.168661 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:08.668586 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:09.171001 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:09.668697 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:10.169866 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:10.668914 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:11.169381 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:11.669187 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:12.168910 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:12.668954 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:13.169251 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:13.669022 18312 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0828 16:55:14.168915 18312 kapi.go:107] duration metric: took 2m31.503163792s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0828 16:55:14.170600 18312 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-001161 cluster.
I0828 16:55:14.171919 18312 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0828 16:55:14.173289 18312 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0828 16:55:14.174604 18312 out.go:177] * Enabled addons: ingress-dns, storage-provisioner, nvidia-device-plugin, default-storageclass, volcano, cloud-spanner, metrics-server, helm-tiller, inspektor-gadget, yakd, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0828 16:55:14.175756 18312 addons.go:510] duration metric: took 2m46.97128152s for enable addons: enabled=[ingress-dns storage-provisioner nvidia-device-plugin default-storageclass volcano cloud-spanner metrics-server helm-tiller inspektor-gadget yakd volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0828 16:55:14.175803 18312 start.go:246] waiting for cluster config update ...
I0828 16:55:14.175829 18312 start.go:255] writing updated cluster config ...
I0828 16:55:14.176128 18312 ssh_runner.go:195] Run: rm -f paused
I0828 16:55:14.222685 18312 start.go:600] kubectl: 1.31.0, cluster: 1.31.0 (minor skew: 0)
I0828 16:55:14.224748 18312 out.go:177] * Done! kubectl is now configured to use "addons-001161" cluster and "default" namespace by default
==> Docker <==
Aug 28 17:04:43 addons-001161 dockerd[1338]: time="2024-08-28T17:04:43.038696578Z" level=info msg="ignoring event" container=4fbe0887114c04dd48281fb8d2d5e123fca1f82ef2e0c43658df8ff11dac34c4 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:43 addons-001161 dockerd[1338]: time="2024-08-28T17:04:43.038755608Z" level=info msg="ignoring event" container=060971608d0ea20cde0d638d77f0d9d37d0ee72db153ba85aac92279e26be1d9 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:43 addons-001161 dockerd[1338]: time="2024-08-28T17:04:43.038787424Z" level=info msg="ignoring event" container=2dd99b4a77bd39129b26d913c44f5a69fe7083dd7590a6dd2ba54e9deb8c71f1 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:43 addons-001161 dockerd[1338]: time="2024-08-28T17:04:43.049223594Z" level=info msg="ignoring event" container=92a41bf80b7499ed7ea3431e685592ddb245707bd2cde3819f16a23b13bf405c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:43 addons-001161 dockerd[1338]: time="2024-08-28T17:04:43.129557989Z" level=info msg="ignoring event" container=f7954d3a864b8fb04e3c7dc431863da57506d2eadb2e38b12082d473e6bcc35e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:43 addons-001161 dockerd[1338]: time="2024-08-28T17:04:43.131471649Z" level=info msg="ignoring event" container=e2cd6d2a11b8ff4b5393b6d81c6b824de263262fefe84553ff62c0922b46c415 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:43 addons-001161 cri-dockerd[1603]: time="2024-08-28T17:04:43Z" level=info msg="Stop pulling image docker.io/kicbase/echo-server:1.0: Status: Downloaded newer image for kicbase/echo-server:1.0"
Aug 28 17:04:43 addons-001161 dockerd[1338]: time="2024-08-28T17:04:43.337071781Z" level=info msg="ignoring event" container=8bf6cd73f33de0b245d31512bc2b4dc3ba1cda6d6afd23d59d1ecabd904da0dd module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:43 addons-001161 dockerd[1338]: time="2024-08-28T17:04:43.367641935Z" level=info msg="ignoring event" container=2ae867b7ac0d5c0e110271aa3a0bf0c3c016c61eca5d6c43a047325b093985cd module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:43 addons-001161 dockerd[1338]: time="2024-08-28T17:04:43.436623758Z" level=info msg="ignoring event" container=d485fd9a06382332a8cdb71fe36c305dea03817091ccaf20ddb636b4267b6250 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:45 addons-001161 dockerd[1338]: time="2024-08-28T17:04:45.562206787Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=d5f3a3e4da5349e1c0b80ea41d6e7d0dfdaec99908fc86b983e4fe11d91dfc10
Aug 28 17:04:45 addons-001161 dockerd[1338]: time="2024-08-28T17:04:45.621588538Z" level=info msg="ignoring event" container=d5f3a3e4da5349e1c0b80ea41d6e7d0dfdaec99908fc86b983e4fe11d91dfc10 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:45 addons-001161 dockerd[1338]: time="2024-08-28T17:04:45.756011911Z" level=info msg="ignoring event" container=cac4fc9b82d2756b07f82a870c8f9a6313e4adc2566095817b469c4b96e85bf7 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:49 addons-001161 dockerd[1338]: time="2024-08-28T17:04:49.286061307Z" level=info msg="ignoring event" container=61d7f34e53eb59d1690f7aff4ad10eb52357636fbbc62a65675425e2228356ff module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:49 addons-001161 dockerd[1338]: time="2024-08-28T17:04:49.287923984Z" level=info msg="ignoring event" container=2893e301323268158c17cdaac493534f81aa0f89d3312dd6f5c2f478c97d5f51 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:49 addons-001161 dockerd[1338]: time="2024-08-28T17:04:49.452536791Z" level=info msg="ignoring event" container=357e42963b6d93bdf7b66a179056a605ca211a33e4f623d93c82056625c663d0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:49 addons-001161 dockerd[1338]: time="2024-08-28T17:04:49.480506919Z" level=info msg="ignoring event" container=69a6f044ad56f8387d8478404ab6d2188b62a44f1856a7dfdc9cfb3419b454c5 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:04:49 addons-001161 dockerd[1338]: time="2024-08-28T17:04:49.902902269Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Aug 28 17:04:49 addons-001161 dockerd[1338]: time="2024-08-28T17:04:49.905017591Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Aug 28 17:05:05 addons-001161 dockerd[1338]: time="2024-08-28T17:05:05.444106855Z" level=info msg="ignoring event" container=8d0fbaa0e73fec2d0a4015bf805cf857f04b20cf3a855a25d477e1f43e339086 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:05:05 addons-001161 dockerd[1338]: time="2024-08-28T17:05:05.942545523Z" level=info msg="ignoring event" container=90a34c1ee610ad22fbf7d7de3440567dcfa5454a1ac01af3503d54f26fc06d0f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:05:05 addons-001161 dockerd[1338]: time="2024-08-28T17:05:05.994931971Z" level=info msg="ignoring event" container=f0e1ba290836fa721f030a0e778f57c5103d3d9c03a9f4c4c02849848a58aabf module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:05:06 addons-001161 dockerd[1338]: time="2024-08-28T17:05:06.070220085Z" level=info msg="ignoring event" container=a1663d10de925c9d42b97682a206d306929c3dfbea55d9c9112dc4da50551477 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 28 17:05:06 addons-001161 cri-dockerd[1603]: time="2024-08-28T17:05:06Z" level=info msg="Failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod \"registry-proxy-vxpf8_kube-system\": unexpected command output nsenter: cannot open /proc/4134/ns/net: No such file or directory\n with error: exit status 1"
Aug 28 17:05:06 addons-001161 dockerd[1338]: time="2024-08-28T17:05:06.148539362Z" level=info msg="ignoring event" container=7a5f997e3619e9e79f01f5f9b2444a38c2b3e5a93564281fd930eee9c6c1ab44 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
4d99622cc0356 kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 23 seconds ago Running hello-world-app 0 92049957cf37e hello-world-app-55bf9c44b4-cbz2c
2c69c99943a1f nginx@sha256:c04c18adc2a407740a397c8407c011fc6c90026a9b65cceddef7ae5484360158 33 seconds ago Running nginx 0 9eecf02e02ec4 nginx
3786a03c152ee ghcr.io/headlamp-k8s/headlamp@sha256:899d106eeb55b0afc4ee6e51c03bc4418de0bd0e79c39744d4d0d751aae6a971 38 seconds ago Running headlamp 0 a2d47450120ee headlamp-57fb76fcdb-xdbsv
1078394cfb7f6 gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 56c786592ee9f gcp-auth-89d5ffd79-nj9gd
a63568559a339 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited patch 0 fefcd2cb40655 ingress-nginx-admission-patch-tljdw
ce3de90cd4d8b registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 20e31ba861659 ingress-nginx-admission-create-dbl8c
f0e1ba290836f gcr.io/k8s-minikube/kube-registry-proxy@sha256:b3fa0b2df8737fdb85ad5918a7e2652527463e357afff83a5e5bb966bcedc367 11 minutes ago Exited registry-proxy 0 7a5f997e3619e registry-proxy-vxpf8
87417b79ce055 rancher/local-path-provisioner@sha256:e34c88ae0affb1cdefbb874140d6339d4a27ec4ee420ae8199cd839997b05246 12 minutes ago Running local-path-provisioner 0 2d0b97a5ce326 local-path-provisioner-86d989889c-26xn4
e2dd2dfda5386 6e38f40d628db 12 minutes ago Running storage-provisioner 0 2c280dcbfeec8 storage-provisioner
2b4552fd3e282 cbb01a7bd410d 12 minutes ago Running coredns 0 dbe599d33e8ac coredns-6f6b679f8f-ls4v7
076884d4e428f ad83b2ca7b09e 12 minutes ago Running kube-proxy 0 5410b6c6efd46 kube-proxy-672sn
fa572e9c58bf3 045733566833c 12 minutes ago Running kube-controller-manager 0 853bed8f59e0a kube-controller-manager-addons-001161
cfd8abd97f381 1766f54c897f0 12 minutes ago Running kube-scheduler 0 f0700b1fbd97d kube-scheduler-addons-001161
b7b2e4a3330fe 604f5db92eaa8 12 minutes ago Running kube-apiserver 0 866c112631ebc kube-apiserver-addons-001161
42d3ba0b25560 2e96e5913fc06 12 minutes ago Running etcd 0 24819c226916a etcd-addons-001161
==> coredns [2b4552fd3e28] <==
[INFO] 10.244.0.22:52678 - 47238 "A IN hello-world-app.default.svc.cluster.local.europe-west1-b.c.k8s-minikube.internal. udp 98 false 512" NXDOMAIN qr,rd,ra 98 0.004936902s
[INFO] 10.244.0.22:42417 - 33339 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.005117738s
[INFO] 10.244.0.22:59803 - 62031 "A IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.004061911s
[INFO] 10.244.0.22:59803 - 50900 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.003997042s
[INFO] 10.244.0.22:52210 - 37327 "AAAA IN hello-world-app.default.svc.cluster.local.europe-west1-b.c.k8s-minikube.internal. udp 98 false 512" NXDOMAIN qr,rd,ra 98 0.004921255s
[INFO] 10.244.0.22:42417 - 57794 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004933005s
[INFO] 10.244.0.22:52678 - 7588 "AAAA IN hello-world-app.default.svc.cluster.local.europe-west1-b.c.k8s-minikube.internal. udp 98 false 512" NXDOMAIN qr,rd,ra 98 0.005130256s
[INFO] 10.244.0.22:58223 - 55087 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005468976s
[INFO] 10.244.0.22:59803 - 51922 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004155905s
[INFO] 10.244.0.22:59803 - 21140 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.065653984s
[INFO] 10.244.0.22:42417 - 20401 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.069507248s
[INFO] 10.244.0.22:58223 - 52656 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.069527699s
[INFO] 10.244.0.22:52210 - 39029 "A IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.069577121s
[INFO] 10.244.0.22:42417 - 49598 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000065397s
[INFO] 10.244.0.22:58223 - 37792 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000047885s
[INFO] 10.244.0.22:59803 - 32288 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000054192s
[INFO] 10.244.0.22:52678 - 30287 "A IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.004929611s
[INFO] 10.244.0.22:52210 - 12719 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.004465517s
[INFO] 10.244.0.22:52678 - 51298 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.004207563s
[INFO] 10.244.0.22:52210 - 49285 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004491648s
[INFO] 10.244.0.22:52678 - 49791 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004019842s
[INFO] 10.244.0.22:52678 - 42379 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.000834487s
[INFO] 10.244.0.22:52210 - 19 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004331623s
[INFO] 10.244.0.22:52678 - 50076 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000069921s
[INFO] 10.244.0.22:52210 - 60718 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000051802s
==> describe nodes <==
Name: addons-001161
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-001161
kubernetes.io/os=linux
minikube.k8s.io/commit=6f256f0bf490fd67de29a75a245d072e85b1b216
minikube.k8s.io/name=addons-001161
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_08_28T16_52_22_0700
minikube.k8s.io/version=v1.33.1
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-001161
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Wed, 28 Aug 2024 16:52:19 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-001161
AcquireTime: <unset>
RenewTime: Wed, 28 Aug 2024 17:04:56 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Wed, 28 Aug 2024 17:04:55 +0000 Wed, 28 Aug 2024 16:52:18 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Wed, 28 Aug 2024 17:04:55 +0000 Wed, 28 Aug 2024 16:52:18 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Wed, 28 Aug 2024 17:04:55 +0000 Wed, 28 Aug 2024 16:52:18 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Wed, 28 Aug 2024 17:04:55 +0000 Wed, 28 Aug 2024 16:52:20 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-001161
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859312Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859312Ki
pods: 110
System Info:
Machine ID: 6339551e14054e8f9b898839bde54d23
System UUID: 6ab449c6-1ecd-4a4c-a6d1-19b548de808c
Boot ID: 7fe637ae-ca35-4a83-9c8d-649da282c777
Kernel Version: 5.15.0-1067-gcp
OS Image: Ubuntu 22.04.4 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.1.2
Kubelet Version: v1.31.0
Kube-Proxy Version:
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (13 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m13s
default hello-world-app-55bf9c44b4-cbz2c 0 (0%) 0 (0%) 0 (0%) 0 (0%) 26s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 36s
gcp-auth gcp-auth-89d5ffd79-nj9gd 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
headlamp headlamp-57fb76fcdb-xdbsv 0 (0%) 0 (0%) 0 (0%) 0 (0%) 41s
kube-system coredns-6f6b679f8f-ls4v7 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-001161 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-001161 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-001161 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-672sn 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-001161 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
local-path-storage local-path-provisioner-86d989889c-26xn4 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal NodeHasSufficientMemory 12m (x8 over 12m) kubelet Node addons-001161 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m (x7 over 12m) kubelet Node addons-001161 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m (x7 over 12m) kubelet Node addons-001161 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-001161 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-001161 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-001161 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-001161 event: Registered Node addons-001161 in Controller
==> dmesg <==
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 5a 64 07 be 82 a9 08 06
[ +2.858248] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 36 b3 3b 7e 6d bf 08 06
[ +6.178629] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff ae 81 aa 42 80 fe 08 06
[ +0.101626] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff ba 4e 70 82 b9 f7 08 06
[ +0.038741] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff ea 99 08 09 8f 5f 08 06
[ +10.461577] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 6a 61 cd 75 c8 af 08 06
[Aug28 16:54] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 56 4a 8e 47 9c 29 08 06
[ +0.016402] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 66 1f 6d a3 89 7c 08 06
[Aug28 16:55] IPv4: martian source 10.244.0.1 from 10.244.0.26, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff be ce 62 93 81 be 08 06
[ +0.000467] IPv4: martian source 10.244.0.26 from 10.244.0.3, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff 4e 08 b0 b4 14 cb 08 06
[Aug28 17:04] IPv4: martian source 10.244.0.1 from 10.244.0.33, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff da a7 fc 21 53 2a 08 06
[ +30.527209] IPv4: martian source 10.244.0.36 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 6a 61 cd 75 c8 af 08 06
[ +0.862289] IPv4: martian source 10.244.0.22 from 10.244.0.3, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 4e 08 b0 b4 14 cb 08 06
==> etcd [42d3ba0b2556] <==
{"level":"info","ts":"2024-08-28T16:52:18.540883Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-08-28T16:52:18.541902Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-001161 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-08-28T16:52:18.541963Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-08-28T16:52:18.541915Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-08-28T16:52:18.542047Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-08-28T16:52:18.542209Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-08-28T16:52:18.542237Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-08-28T16:52:18.542737Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-08-28T16:52:18.542890Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-08-28T16:52:18.542979Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-08-28T16:52:18.543176Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-08-28T16:52:18.543316Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-08-28T16:52:18.543973Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-08-28T16:52:18.544505Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-08-28T16:52:31.329313Z","caller":"traceutil/trace.go:171","msg":"trace[969535432] transaction","detail":"{read_only:false; response_revision:441; number_of_response:1; }","duration":"175.387583ms","start":"2024-08-28T16:52:31.153903Z","end":"2024-08-28T16:52:31.329291Z","steps":["trace[969535432] 'process raft request' (duration: 88.205168ms)","trace[969535432] 'compare' (duration: 86.551019ms)"],"step_count":2}
{"level":"info","ts":"2024-08-28T16:52:31.329466Z","caller":"traceutil/trace.go:171","msg":"trace[1760552887] transaction","detail":"{read_only:false; response_revision:443; number_of_response:1; }","duration":"100.266999ms","start":"2024-08-28T16:52:31.229174Z","end":"2024-08-28T16:52:31.329441Z","steps":["trace[1760552887] 'process raft request' (duration: 99.756205ms)"],"step_count":1}
{"level":"info","ts":"2024-08-28T16:52:31.329643Z","caller":"traceutil/trace.go:171","msg":"trace[151707045] transaction","detail":"{read_only:false; response_revision:442; number_of_response:1; }","duration":"175.116783ms","start":"2024-08-28T16:52:31.154516Z","end":"2024-08-28T16:52:31.329633Z","steps":["trace[151707045] 'process raft request' (duration: 174.29345ms)"],"step_count":1}
{"level":"info","ts":"2024-08-28T16:52:33.431709Z","caller":"traceutil/trace.go:171","msg":"trace[740143208] transaction","detail":"{read_only:false; response_revision:555; number_of_response:1; }","duration":"101.373007ms","start":"2024-08-28T16:52:33.330319Z","end":"2024-08-28T16:52:33.431692Z","steps":["trace[740143208] 'process raft request' (duration: 100.783849ms)"],"step_count":1}
{"level":"info","ts":"2024-08-28T16:52:33.431960Z","caller":"traceutil/trace.go:171","msg":"trace[3498891] transaction","detail":"{read_only:false; response_revision:556; number_of_response:1; }","duration":"101.529895ms","start":"2024-08-28T16:52:33.330419Z","end":"2024-08-28T16:52:33.431948Z","steps":["trace[3498891] 'process raft request' (duration: 100.945865ms)"],"step_count":1}
{"level":"info","ts":"2024-08-28T16:52:38.741644Z","caller":"traceutil/trace.go:171","msg":"trace[123362413] transaction","detail":"{read_only:false; response_revision:824; number_of_response:1; }","duration":"111.426096ms","start":"2024-08-28T16:52:38.630196Z","end":"2024-08-28T16:52:38.741622Z","steps":["trace[123362413] 'process raft request' (duration: 111.080397ms)"],"step_count":1}
{"level":"warn","ts":"2024-08-28T16:53:08.977033Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"127.254113ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-08-28T16:53:08.977130Z","caller":"traceutil/trace.go:171","msg":"trace[852628635] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:1093; }","duration":"127.369851ms","start":"2024-08-28T16:53:08.849746Z","end":"2024-08-28T16:53:08.977116Z","steps":["trace[852628635] 'range keys from in-memory index tree' (duration: 127.202923ms)"],"step_count":1}
{"level":"info","ts":"2024-08-28T17:02:18.561543Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1917}
{"level":"info","ts":"2024-08-28T17:02:18.585868Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1917,"took":"23.775439ms","hash":3470285451,"current-db-size-bytes":8998912,"current-db-size":"9.0 MB","current-db-size-in-use-bytes":5083136,"current-db-size-in-use":"5.1 MB"}
{"level":"info","ts":"2024-08-28T17:02:18.585910Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":3470285451,"revision":1917,"compact-revision":-1}
==> gcp-auth [1078394cfb7f] <==
2024/08/28 16:55:53 Ready to write response ...
2024/08/28 17:03:55 Ready to marshal response ...
2024/08/28 17:03:55 Ready to write response ...
2024/08/28 17:03:55 Ready to marshal response ...
2024/08/28 17:03:55 Ready to write response ...
2024/08/28 17:04:03 Ready to marshal response ...
2024/08/28 17:04:03 Ready to write response ...
2024/08/28 17:04:05 Ready to marshal response ...
2024/08/28 17:04:05 Ready to write response ...
2024/08/28 17:04:06 Ready to marshal response ...
2024/08/28 17:04:06 Ready to write response ...
2024/08/28 17:04:07 Ready to marshal response ...
2024/08/28 17:04:07 Ready to write response ...
2024/08/28 17:04:25 Ready to marshal response ...
2024/08/28 17:04:25 Ready to write response ...
2024/08/28 17:04:25 Ready to marshal response ...
2024/08/28 17:04:25 Ready to write response ...
2024/08/28 17:04:25 Ready to marshal response ...
2024/08/28 17:04:25 Ready to write response ...
2024/08/28 17:04:30 Ready to marshal response ...
2024/08/28 17:04:30 Ready to write response ...
2024/08/28 17:04:30 Ready to marshal response ...
2024/08/28 17:04:30 Ready to write response ...
2024/08/28 17:04:40 Ready to marshal response ...
2024/08/28 17:04:40 Ready to write response ...
==> kernel <==
17:05:06 up 47 min, 0 users, load average: 0.30, 0.36, 0.33
Linux addons-001161 5.15.0-1067-gcp #75~20.04.1-Ubuntu SMP Wed Aug 7 20:43:22 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.4 LTS"
==> kube-apiserver [b7b2e4a3330f] <==
W0828 16:55:44.749258 1 cacher.go:171] Terminating all watchers from cacher queues.scheduling.volcano.sh
W0828 16:55:45.046281 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0828 16:55:45.344133 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0828 16:55:45.607814 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0828 17:04:06.671312 1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
I0828 17:04:12.572143 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0828 17:04:24.348707 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
I0828 17:04:25.167468 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.102.85.232"}
W0828 17:04:25.363784 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0828 17:04:29.905036 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0828 17:04:30.070935 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.109.68.36"}
I0828 17:04:40.547264 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.110.70.195"}
I0828 17:04:49.193532 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0828 17:04:49.193584 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0828 17:04:49.204962 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0828 17:04:49.205019 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0828 17:04:49.206415 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0828 17:04:49.206455 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0828 17:04:49.218072 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0828 17:04:49.218108 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0828 17:04:49.236878 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0828 17:04:49.236918 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0828 17:04:50.206691 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W0828 17:04:50.237558 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
W0828 17:04:50.334559 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
==> kube-controller-manager [fa572e9c58bf] <==
I0828 17:04:52.558135 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="ingress-nginx"
W0828 17:04:53.280651 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0828 17:04:53.280688 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0828 17:04:53.613149 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0828 17:04:53.613194 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0828 17:04:53.802876 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0828 17:04:53.802914 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0828 17:04:53.832971 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0828 17:04:53.833013 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0828 17:04:55.568629 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-001161"
I0828 17:04:56.876759 1 shared_informer.go:313] Waiting for caches to sync for resource quota
I0828 17:04:56.876792 1 shared_informer.go:320] Caches are synced for resource quota
I0828 17:04:57.190159 1 shared_informer.go:313] Waiting for caches to sync for garbage collector
I0828 17:04:57.190203 1 shared_informer.go:320] Caches are synced for garbage collector
W0828 17:04:57.637444 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0828 17:04:57.637486 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0828 17:04:57.880157 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0828 17:04:57.880193 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0828 17:04:57.887626 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0828 17:04:57.887671 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0828 17:04:58.461435 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0828 17:04:58.461472 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0828 17:04:59.978794 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0828 17:04:59.978828 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0828 17:05:05.895697 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-6fb4cdfc84" duration="15.209µs"
==> kube-proxy [076884d4e428] <==
I0828 16:52:29.842922 1 server_linux.go:66] "Using iptables proxy"
I0828 16:52:30.448753 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0828 16:52:30.448843 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0828 16:52:31.034301 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0828 16:52:31.034374 1 server_linux.go:169] "Using iptables Proxier"
I0828 16:52:31.048885 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0828 16:52:31.049198 1 server.go:483] "Version info" version="v1.31.0"
I0828 16:52:31.049220 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0828 16:52:31.051232 1 config.go:197] "Starting service config controller"
I0828 16:52:31.051245 1 shared_informer.go:313] Waiting for caches to sync for service config
I0828 16:52:31.051267 1 config.go:104] "Starting endpoint slice config controller"
I0828 16:52:31.051272 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0828 16:52:31.051298 1 config.go:326] "Starting node config controller"
I0828 16:52:31.051303 1 shared_informer.go:313] Waiting for caches to sync for node config
I0828 16:52:31.151707 1 shared_informer.go:320] Caches are synced for node config
I0828 16:52:31.151753 1 shared_informer.go:320] Caches are synced for service config
I0828 16:52:31.151782 1 shared_informer.go:320] Caches are synced for endpoint slice config
==> kube-scheduler [cfd8abd97f38] <==
E0828 16:52:19.741304 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
E0828 16:52:19.740840 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0828 16:52:19.741591 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
W0828 16:52:19.741624 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0828 16:52:19.741633 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0828 16:52:19.741593 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0828 16:52:19.741656 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0828 16:52:19.741665 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0828 16:52:19.741782 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0828 16:52:19.741819 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0828 16:52:20.549606 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0828 16:52:20.549657 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0828 16:52:20.585902 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0828 16:52:20.585936 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0828 16:52:20.590110 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0828 16:52:20.590155 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0828 16:52:20.601470 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0828 16:52:20.601513 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0828 16:52:20.682170 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0828 16:52:20.682219 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0828 16:52:20.717408 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0828 16:52:20.717452 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0828 16:52:20.807052 1 reflector.go:561] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0828 16:52:20.807091 1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
I0828 16:52:23.434278 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Aug 28 17:04:49 addons-001161 kubelet[2454]: E0828 17:04:49.905418 2454 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = Error response from daemon: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed" image="gcr.io/k8s-minikube/busybox:latest"
Aug 28 17:04:49 addons-001161 kubelet[2454]: E0828 17:04:49.905571 2454 kuberuntime_manager.go:1272] "Unhandled Error" err="container &Container{Name:registry-test,Image:gcr.io/k8s-minikube/busybox,Command:[],Args:[sh -c wget --spider -S http://registry.kube-system.svc.cluster.local],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:GOOGLE_APPLICATION_CREDENTIALS,Value:/google-app-creds.json,ValueFrom:nil,},EnvVar{Name:PROJECT_ID,Value:this_is_fake,ValueFrom:nil,},EnvVar{Name:GCP_PROJECT,Value:this_is_fake,ValueFrom:nil,},EnvVar{Name:GCLOUD_PROJECT,Value:this_is_fake,ValueFrom:nil,},EnvVar{Name:GOOGLE_CLOUD_PROJECT,Value:this_is_fake,ValueFrom:nil,},EnvVar{Name:CLOUDSDK_CORE_PROJECT,Value:this_is_fake,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4hr5w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,Su
bPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:gcp-creds,ReadOnly:true,MountPath:/google-app-creds.json,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:true,StdinOnce:true,TTY:true,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod registry-test_default(dadb66fd-da22-4715-b324-79f38bac1bb1): ErrImagePull: Error response from daemon: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed" logger="UnhandledError"
Aug 28 17:04:49 addons-001161 kubelet[2454]: E0828 17:04:49.906771 2454 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ErrImagePull: \"Error response from daemon: Head \\\"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\\\": unauthorized: authentication failed\"" pod="default/registry-test" podUID="dadb66fd-da22-4715-b324-79f38bac1bb1"
Aug 28 17:04:49 addons-001161 kubelet[2454]: I0828 17:04:49.912108 2454 scope.go:117] "RemoveContainer" containerID="2893e301323268158c17cdaac493534f81aa0f89d3312dd6f5c2f478c97d5f51"
Aug 28 17:04:49 addons-001161 kubelet[2454]: E0828 17:04:49.912881 2454 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 2893e301323268158c17cdaac493534f81aa0f89d3312dd6f5c2f478c97d5f51" containerID="2893e301323268158c17cdaac493534f81aa0f89d3312dd6f5c2f478c97d5f51"
Aug 28 17:04:49 addons-001161 kubelet[2454]: I0828 17:04:49.912928 2454 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"2893e301323268158c17cdaac493534f81aa0f89d3312dd6f5c2f478c97d5f51"} err="failed to get container status \"2893e301323268158c17cdaac493534f81aa0f89d3312dd6f5c2f478c97d5f51\": rpc error: code = Unknown desc = Error response from daemon: No such container: 2893e301323268158c17cdaac493534f81aa0f89d3312dd6f5c2f478c97d5f51"
Aug 28 17:04:50 addons-001161 kubelet[2454]: E0828 17:04:50.777267 2454 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="21d6da0f-4e2a-48c8-9164-342017356226"
Aug 28 17:04:51 addons-001161 kubelet[2454]: I0828 17:04:51.782960 2454 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bfc7192-ef8b-4ccf-87f1-451168ab8fb6" path="/var/lib/kubelet/pods/4bfc7192-ef8b-4ccf-87f1-451168ab8fb6/volumes"
Aug 28 17:04:51 addons-001161 kubelet[2454]: I0828 17:04:51.783311 2454 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="794b01fa-aedf-4cc2-89b1-a6f1ad17e165" path="/var/lib/kubelet/pods/794b01fa-aedf-4cc2-89b1-a6f1ad17e165/volumes"
Aug 28 17:05:02 addons-001161 kubelet[2454]: E0828 17:05:02.777288 2454 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox\\\"\"" pod="default/registry-test" podUID="dadb66fd-da22-4715-b324-79f38bac1bb1"
Aug 28 17:05:03 addons-001161 kubelet[2454]: E0828 17:05:03.776813 2454 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="21d6da0f-4e2a-48c8-9164-342017356226"
Aug 28 17:05:04 addons-001161 kubelet[2454]: I0828 17:05:04.775415 2454 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="kube-system/registry-6fb4cdfc84-lbgl8" secret="" err="secret \"gcp-auth\" not found"
Aug 28 17:05:05 addons-001161 kubelet[2454]: I0828 17:05:05.589713 2454 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/dadb66fd-da22-4715-b324-79f38bac1bb1-gcp-creds\") pod \"dadb66fd-da22-4715-b324-79f38bac1bb1\" (UID: \"dadb66fd-da22-4715-b324-79f38bac1bb1\") "
Aug 28 17:05:05 addons-001161 kubelet[2454]: I0828 17:05:05.589774 2454 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hr5w\" (UniqueName: \"kubernetes.io/projected/dadb66fd-da22-4715-b324-79f38bac1bb1-kube-api-access-4hr5w\") pod \"dadb66fd-da22-4715-b324-79f38bac1bb1\" (UID: \"dadb66fd-da22-4715-b324-79f38bac1bb1\") "
Aug 28 17:05:05 addons-001161 kubelet[2454]: I0828 17:05:05.589821 2454 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dadb66fd-da22-4715-b324-79f38bac1bb1-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "dadb66fd-da22-4715-b324-79f38bac1bb1" (UID: "dadb66fd-da22-4715-b324-79f38bac1bb1"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Aug 28 17:05:05 addons-001161 kubelet[2454]: I0828 17:05:05.599208 2454 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dadb66fd-da22-4715-b324-79f38bac1bb1-kube-api-access-4hr5w" (OuterVolumeSpecName: "kube-api-access-4hr5w") pod "dadb66fd-da22-4715-b324-79f38bac1bb1" (UID: "dadb66fd-da22-4715-b324-79f38bac1bb1"). InnerVolumeSpecName "kube-api-access-4hr5w". PluginName "kubernetes.io/projected", VolumeGidValue ""
Aug 28 17:05:05 addons-001161 kubelet[2454]: I0828 17:05:05.690544 2454 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-4hr5w\" (UniqueName: \"kubernetes.io/projected/dadb66fd-da22-4715-b324-79f38bac1bb1-kube-api-access-4hr5w\") on node \"addons-001161\" DevicePath \"\""
Aug 28 17:05:05 addons-001161 kubelet[2454]: I0828 17:05:05.690584 2454 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/dadb66fd-da22-4715-b324-79f38bac1bb1-gcp-creds\") on node \"addons-001161\" DevicePath \"\""
Aug 28 17:05:06 addons-001161 kubelet[2454]: I0828 17:05:06.151276 2454 scope.go:117] "RemoveContainer" containerID="90a34c1ee610ad22fbf7d7de3440567dcfa5454a1ac01af3503d54f26fc06d0f"
Aug 28 17:05:06 addons-001161 kubelet[2454]: I0828 17:05:06.294557 2454 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khll9\" (UniqueName: \"kubernetes.io/projected/243f93fd-9335-4ba8-a33c-d9ed0affd92f-kube-api-access-khll9\") pod \"243f93fd-9335-4ba8-a33c-d9ed0affd92f\" (UID: \"243f93fd-9335-4ba8-a33c-d9ed0affd92f\") "
Aug 28 17:05:06 addons-001161 kubelet[2454]: I0828 17:05:06.294600 2454 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khglz\" (UniqueName: \"kubernetes.io/projected/6d3d6e6b-090e-4033-9117-635faf64b5bf-kube-api-access-khglz\") pod \"6d3d6e6b-090e-4033-9117-635faf64b5bf\" (UID: \"6d3d6e6b-090e-4033-9117-635faf64b5bf\") "
Aug 28 17:05:06 addons-001161 kubelet[2454]: I0828 17:05:06.296398 2454 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d3d6e6b-090e-4033-9117-635faf64b5bf-kube-api-access-khglz" (OuterVolumeSpecName: "kube-api-access-khglz") pod "6d3d6e6b-090e-4033-9117-635faf64b5bf" (UID: "6d3d6e6b-090e-4033-9117-635faf64b5bf"). InnerVolumeSpecName "kube-api-access-khglz". PluginName "kubernetes.io/projected", VolumeGidValue ""
Aug 28 17:05:06 addons-001161 kubelet[2454]: I0828 17:05:06.296466 2454 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/243f93fd-9335-4ba8-a33c-d9ed0affd92f-kube-api-access-khll9" (OuterVolumeSpecName: "kube-api-access-khll9") pod "243f93fd-9335-4ba8-a33c-d9ed0affd92f" (UID: "243f93fd-9335-4ba8-a33c-d9ed0affd92f"). InnerVolumeSpecName "kube-api-access-khll9". PluginName "kubernetes.io/projected", VolumeGidValue ""
Aug 28 17:05:06 addons-001161 kubelet[2454]: I0828 17:05:06.395799 2454 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-khll9\" (UniqueName: \"kubernetes.io/projected/243f93fd-9335-4ba8-a33c-d9ed0affd92f-kube-api-access-khll9\") on node \"addons-001161\" DevicePath \"\""
Aug 28 17:05:06 addons-001161 kubelet[2454]: I0828 17:05:06.395841 2454 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-khglz\" (UniqueName: \"kubernetes.io/projected/6d3d6e6b-090e-4033-9117-635faf64b5bf-kube-api-access-khglz\") on node \"addons-001161\" DevicePath \"\""
==> storage-provisioner [e2dd2dfda538] <==
I0828 16:52:35.628854 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0828 16:52:35.649548 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0828 16:52:35.649617 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0828 16:52:35.735473 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0828 16:52:35.735662 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-001161_e1208ea2-6641-4bed-bf50-0895c9737b36!
I0828 16:52:35.736821 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"d66681f0-070c-4cd7-95e9-2c3817205a70", APIVersion:"v1", ResourceVersion:"652", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-001161_e1208ea2-6641-4bed-bf50-0895c9737b36 became leader
I0828 16:52:35.835744 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-001161_e1208ea2-6641-4bed-bf50-0895c9737b36!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-001161 -n addons-001161
helpers_test.go:261: (dbg) Run: kubectl --context addons-001161 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-001161 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-001161 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-001161/192.168.49.2
Start Time: Wed, 28 Aug 2024 16:55:53 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.28
IPs:
IP: 10.244.0.28
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-vwmsp (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-vwmsp:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m14s default-scheduler Successfully assigned default/busybox to addons-001161
Normal Pulling 7m55s (x4 over 9m14s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m55s (x4 over 9m14s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m55s (x4 over 9m14s) kubelet Error: ErrImagePull
Warning Failed 7m32s (x6 over 9m13s) kubelet Error: ImagePullBackOff
Normal BackOff 4m5s (x21 over 9m13s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (72.41s)