=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:332: registry stabilized in 2.33019ms
addons_test.go:334: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-6fb4cdfc84-7xs8r" [0bc8f454-eced-450f-ab5d-b961648307b9] Running
addons_test.go:334: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.003230347s
addons_test.go:337: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-wwxwv" [ef74c371-cbb9-4c81-91dd-dcbc748f81d0] Running
addons_test.go:337: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.003548486s
addons_test.go:342: (dbg) Run: kubectl --context addons-505336 delete po -l run=registry-test --now
addons_test.go:347: (dbg) Run: kubectl --context addons-505336 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:347: (dbg) Non-zero exit: kubectl --context addons-505336 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.069999559s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:349: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-505336 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:353: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:361: (dbg) Run: out/minikube-linux-amd64 -p addons-505336 ip
2024/08/29 19:09:10 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:390: (dbg) Run: out/minikube-linux-amd64 -p addons-505336 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-505336
helpers_test.go:235: (dbg) docker inspect addons-505336:
-- stdout --
[
{
"Id": "8ef4bbc72da474829e04afb352bc954566bb82aaa67e78c02a0071c69b58d9b1",
"Created": "2024-08-29T18:56:12.889590417Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 427561,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-08-29T18:56:12.99355564Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:cf9874f1e25d62abde3fdda0022141a8ec82ded75077d073b80dc8f90194cf19",
"ResolvConfPath": "/var/lib/docker/containers/8ef4bbc72da474829e04afb352bc954566bb82aaa67e78c02a0071c69b58d9b1/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/8ef4bbc72da474829e04afb352bc954566bb82aaa67e78c02a0071c69b58d9b1/hostname",
"HostsPath": "/var/lib/docker/containers/8ef4bbc72da474829e04afb352bc954566bb82aaa67e78c02a0071c69b58d9b1/hosts",
"LogPath": "/var/lib/docker/containers/8ef4bbc72da474829e04afb352bc954566bb82aaa67e78c02a0071c69b58d9b1/8ef4bbc72da474829e04afb352bc954566bb82aaa67e78c02a0071c69b58d9b1-json.log",
"Name": "/addons-505336",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-505336:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-505336",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/86b0c4497f83061cd04974fda87a0c682723e945a7c6fbd6e5d6e042555617a6-init/diff:/var/lib/docker/overlay2/1f5f6f094bf9cdbb177c01e7ca97214612c2dac25cb3e288a6ba736aeaa2c5c2/diff",
"MergedDir": "/var/lib/docker/overlay2/86b0c4497f83061cd04974fda87a0c682723e945a7c6fbd6e5d6e042555617a6/merged",
"UpperDir": "/var/lib/docker/overlay2/86b0c4497f83061cd04974fda87a0c682723e945a7c6fbd6e5d6e042555617a6/diff",
"WorkDir": "/var/lib/docker/overlay2/86b0c4497f83061cd04974fda87a0c682723e945a7c6fbd6e5d6e042555617a6/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-505336",
"Source": "/var/lib/docker/volumes/addons-505336/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-505336",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-505336",
"name.minikube.sigs.k8s.io": "addons-505336",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "fd17ded43393cc6869ec5fb504a4afde883f25c7a15d9c8499bb1b97097f66ce",
"SandboxKey": "/var/run/docker/netns/fd17ded43393",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32803"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32804"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32807"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32805"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32806"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-505336": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "600cb417e83e89112eef30e4731ed0ea8e8bb05979f28293f72fc192bf2a6925",
"EndpointID": "c84db0bff99942b0798ea7668ee6cd30a454078180afe266de15b1f57ee90ce3",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-505336",
"8ef4bbc72da4"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-505336 -n addons-505336
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-505336 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-docker-211674 | download-docker-211674 | jenkins | v1.33.1 | 29 Aug 24 18:55 UTC | 29 Aug 24 18:55 UTC |
| start | --download-only -p | binary-mirror-967186 | jenkins | v1.33.1 | 29 Aug 24 18:55 UTC | |
| | binary-mirror-967186 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:42019 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-967186 | binary-mirror-967186 | jenkins | v1.33.1 | 29 Aug 24 18:55 UTC | 29 Aug 24 18:55 UTC |
| addons | disable dashboard -p | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 18:55 UTC | |
| | addons-505336 | | | | | |
| addons | enable dashboard -p | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 18:55 UTC | |
| | addons-505336 | | | | | |
| start | -p addons-505336 --wait=true | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 18:55 UTC | 29 Aug 24 18:59 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | addons-505336 addons disable | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 18:59 UTC | 29 Aug 24 18:59 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | addons-505336 addons | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-505336 addons disable | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ssh | addons-505336 ssh cat | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | /opt/local-path-provisioner/pvc-e7cb1702-6246-4f1e-af32-73da81c1bbe3_default_test-pvc/file1 | | | | | |
| addons | addons-505336 addons disable | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-505336 addons disable | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | -p addons-505336 | | | | | |
| addons | disable cloud-spanner -p | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | addons-505336 | | | | | |
| addons | enable headlamp | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | -p addons-505336 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | addons-505336 | | | | | |
| addons | addons-505336 addons disable | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-505336 addons | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-505336 ssh curl -s | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-505336 ip | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| addons | addons-505336 addons disable | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-505336 addons disable | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| addons | addons-505336 addons | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:08 UTC | 29 Aug 24 19:08 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-505336 ip | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:09 UTC | 29 Aug 24 19:09 UTC |
| addons | addons-505336 addons disable | addons-505336 | jenkins | v1.33.1 | 29 Aug 24 19:09 UTC | 29 Aug 24 19:09 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/08/29 18:55:49
Running on machine: ubuntu-20-agent-14
Binary: Built with gc go1.22.5 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0829 18:55:49.472236 426808 out.go:345] Setting OutFile to fd 1 ...
I0829 18:55:49.472455 426808 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0829 18:55:49.472463 426808 out.go:358] Setting ErrFile to fd 2...
I0829 18:55:49.472468 426808 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0829 18:55:49.472611 426808 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19530-418716/.minikube/bin
I0829 18:55:49.473153 426808 out.go:352] Setting JSON to false
I0829 18:55:49.473954 426808 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":77895,"bootTime":1724879854,"procs":172,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1067-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0829 18:55:49.474009 426808 start.go:139] virtualization: kvm guest
I0829 18:55:49.476044 426808 out.go:177] * [addons-505336] minikube v1.33.1 on Ubuntu 20.04 (kvm/amd64)
I0829 18:55:49.477351 426808 notify.go:220] Checking for updates...
I0829 18:55:49.477363 426808 out.go:177] - MINIKUBE_LOCATION=19530
I0829 18:55:49.478639 426808 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0829 18:55:49.479952 426808 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19530-418716/kubeconfig
I0829 18:55:49.481081 426808 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19530-418716/.minikube
I0829 18:55:49.482221 426808 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0829 18:55:49.483408 426808 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0829 18:55:49.484628 426808 driver.go:392] Setting default libvirt URI to qemu:///system
I0829 18:55:49.504861 426808 docker.go:123] docker version: linux-27.2.0:Docker Engine - Community
I0829 18:55:49.505005 426808 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0829 18:55:49.548490 426808 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-08-29 18:55:49.540064972 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1067-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.0 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:472731909fa34bd7bc9c087e4c27943f9835f111 Expected:472731909fa34bd7bc9c087e4c27943f9835f111} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0829 18:55:49.548592 426808 docker.go:307] overlay module found
I0829 18:55:49.550200 426808 out.go:177] * Using the docker driver based on user configuration
I0829 18:55:49.551457 426808 start.go:297] selected driver: docker
I0829 18:55:49.551478 426808 start.go:901] validating driver "docker" against <nil>
I0829 18:55:49.551495 426808 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0829 18:55:49.552437 426808 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0829 18:55:49.598001 426808 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-08-29 18:55:49.589555401 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1067-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.0 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:472731909fa34bd7bc9c087e4c27943f9835f111 Expected:472731909fa34bd7bc9c087e4c27943f9835f111} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0829 18:55:49.598155 426808 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0829 18:55:49.598348 426808 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0829 18:55:49.600079 426808 out.go:177] * Using Docker driver with root privileges
I0829 18:55:49.601426 426808 cni.go:84] Creating CNI manager for ""
I0829 18:55:49.601459 426808 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0829 18:55:49.601478 426808 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0829 18:55:49.601536 426808 start.go:340] cluster config:
{Name:addons-505336 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-505336 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0829 18:55:49.602917 426808 out.go:177] * Starting "addons-505336" primary control-plane node in "addons-505336" cluster
I0829 18:55:49.603988 426808 cache.go:121] Beginning downloading kic base image for docker with docker
I0829 18:55:49.605144 426808 out.go:177] * Pulling base image v0.0.44-1724862063-19530 ...
I0829 18:55:49.606188 426808 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0829 18:55:49.606217 426808 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 in local docker daemon
I0829 18:55:49.606233 426808 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19530-418716/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4
I0829 18:55:49.606242 426808 cache.go:56] Caching tarball of preloaded images
I0829 18:55:49.606337 426808 preload.go:172] Found /home/jenkins/minikube-integration/19530-418716/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0829 18:55:49.606348 426808 cache.go:59] Finished verifying existence of preloaded tar for v1.31.0 on docker
I0829 18:55:49.606658 426808 profile.go:143] Saving config to /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/config.json ...
I0829 18:55:49.606678 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/config.json: {Name:mk5cb17a28b802907616a2b7f8d0dca7d8c84314 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:55:49.621301 426808 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 to local cache
I0829 18:55:49.621406 426808 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 in local cache directory
I0829 18:55:49.621420 426808 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 in local cache directory, skipping pull
I0829 18:55:49.621425 426808 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 exists in cache, skipping pull
I0829 18:55:49.621432 426808 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 as a tarball
I0829 18:55:49.621440 426808 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 from local cache
I0829 18:56:01.263472 426808 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 from cached tarball
I0829 18:56:01.263509 426808 cache.go:194] Successfully downloaded all kic artifacts
I0829 18:56:01.263553 426808 start.go:360] acquireMachinesLock for addons-505336: {Name:mkbfe615c89ee2b0f36152d22e96b58565330b4c Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0829 18:56:01.263648 426808 start.go:364] duration metric: took 74.32µs to acquireMachinesLock for "addons-505336"
I0829 18:56:01.263671 426808 start.go:93] Provisioning new machine with config: &{Name:addons-505336 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-505336 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0829 18:56:01.263755 426808 start.go:125] createHost starting for "" (driver="docker")
I0829 18:56:01.265626 426808 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0829 18:56:01.265855 426808 start.go:159] libmachine.API.Create for "addons-505336" (driver="docker")
I0829 18:56:01.265889 426808 client.go:168] LocalClient.Create starting
I0829 18:56:01.265971 426808 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19530-418716/.minikube/certs/ca.pem
I0829 18:56:01.331464 426808 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19530-418716/.minikube/certs/cert.pem
I0829 18:56:01.723592 426808 cli_runner.go:164] Run: docker network inspect addons-505336 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0829 18:56:01.738388 426808 cli_runner.go:211] docker network inspect addons-505336 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0829 18:56:01.738451 426808 network_create.go:284] running [docker network inspect addons-505336] to gather additional debugging logs...
I0829 18:56:01.738473 426808 cli_runner.go:164] Run: docker network inspect addons-505336
W0829 18:56:01.752651 426808 cli_runner.go:211] docker network inspect addons-505336 returned with exit code 1
I0829 18:56:01.752679 426808 network_create.go:287] error running [docker network inspect addons-505336]: docker network inspect addons-505336: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-505336 not found
I0829 18:56:01.752694 426808 network_create.go:289] output of [docker network inspect addons-505336]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-505336 not found
** /stderr **
I0829 18:56:01.752769 426808 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0829 18:56:01.767534 426808 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc0013588c0}
I0829 18:56:01.767577 426808 network_create.go:124] attempt to create docker network addons-505336 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0829 18:56:01.767623 426808 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-505336 addons-505336
I0829 18:56:01.825407 426808 network_create.go:108] docker network addons-505336 192.168.49.0/24 created
I0829 18:56:01.825437 426808 kic.go:121] calculated static IP "192.168.49.2" for the "addons-505336" container
I0829 18:56:01.825491 426808 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0829 18:56:01.840317 426808 cli_runner.go:164] Run: docker volume create addons-505336 --label name.minikube.sigs.k8s.io=addons-505336 --label created_by.minikube.sigs.k8s.io=true
I0829 18:56:01.855892 426808 oci.go:103] Successfully created a docker volume addons-505336
I0829 18:56:01.855963 426808 cli_runner.go:164] Run: docker run --rm --name addons-505336-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-505336 --entrypoint /usr/bin/test -v addons-505336:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 -d /var/lib
I0829 18:56:08.941593 426808 cli_runner.go:217] Completed: docker run --rm --name addons-505336-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-505336 --entrypoint /usr/bin/test -v addons-505336:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 -d /var/lib: (7.085593529s)
I0829 18:56:08.941623 426808 oci.go:107] Successfully prepared a docker volume addons-505336
I0829 18:56:08.941645 426808 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0829 18:56:08.941671 426808 kic.go:194] Starting extracting preloaded images to volume ...
I0829 18:56:08.941725 426808 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19530-418716/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-505336:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 -I lz4 -xf /preloaded.tar -C /extractDir
I0829 18:56:12.830876 426808 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19530-418716/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-505336:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 -I lz4 -xf /preloaded.tar -C /extractDir: (3.889113932s)
I0829 18:56:12.830913 426808 kic.go:203] duration metric: took 3.889239092s to extract preloaded images to volume ...
W0829 18:56:12.831065 426808 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0829 18:56:12.831178 426808 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0829 18:56:12.872995 426808 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-505336 --name addons-505336 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-505336 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-505336 --network addons-505336 --ip 192.168.49.2 --volume addons-505336:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0
I0829 18:56:13.160986 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Running}}
I0829 18:56:13.177979 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:13.195798 426808 cli_runner.go:164] Run: docker exec addons-505336 stat /var/lib/dpkg/alternatives/iptables
I0829 18:56:13.237707 426808 oci.go:144] the created container "addons-505336" has a running status.
I0829 18:56:13.237740 426808 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa...
I0829 18:56:13.347503 426808 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0829 18:56:13.366565 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:13.383873 426808 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0829 18:56:13.383899 426808 kic_runner.go:114] Args: [docker exec --privileged addons-505336 chown docker:docker /home/docker/.ssh/authorized_keys]
I0829 18:56:13.426772 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:13.447053 426808 machine.go:93] provisionDockerMachine start ...
I0829 18:56:13.447178 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:13.464011 426808 main.go:141] libmachine: Using SSH client type: native
I0829 18:56:13.464239 426808 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32803 <nil> <nil>}
I0829 18:56:13.464254 426808 main.go:141] libmachine: About to run SSH command:
hostname
I0829 18:56:13.464876 426808 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: read tcp 127.0.0.1:38492->127.0.0.1:32803: read: connection reset by peer
I0829 18:56:16.586205 426808 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-505336
I0829 18:56:16.586241 426808 ubuntu.go:169] provisioning hostname "addons-505336"
I0829 18:56:16.586298 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:16.602372 426808 main.go:141] libmachine: Using SSH client type: native
I0829 18:56:16.602569 426808 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32803 <nil> <nil>}
I0829 18:56:16.602584 426808 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-505336 && echo "addons-505336" | sudo tee /etc/hostname
I0829 18:56:16.729321 426808 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-505336
I0829 18:56:16.729390 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:16.745113 426808 main.go:141] libmachine: Using SSH client type: native
I0829 18:56:16.745348 426808 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32803 <nil> <nil>}
I0829 18:56:16.745371 426808 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-505336' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-505336/g' /etc/hosts;
else
echo '127.0.1.1 addons-505336' | sudo tee -a /etc/hosts;
fi
fi
I0829 18:56:16.862621 426808 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0829 18:56:16.862652 426808 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19530-418716/.minikube CaCertPath:/home/jenkins/minikube-integration/19530-418716/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19530-418716/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19530-418716/.minikube}
I0829 18:56:16.862709 426808 ubuntu.go:177] setting up certificates
I0829 18:56:16.862719 426808 provision.go:84] configureAuth start
I0829 18:56:16.862767 426808 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-505336
I0829 18:56:16.878066 426808 provision.go:143] copyHostCerts
I0829 18:56:16.878141 426808 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19530-418716/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19530-418716/.minikube/ca.pem (1078 bytes)
I0829 18:56:16.878287 426808 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19530-418716/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19530-418716/.minikube/cert.pem (1123 bytes)
I0829 18:56:16.878353 426808 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19530-418716/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19530-418716/.minikube/key.pem (1675 bytes)
I0829 18:56:16.878405 426808 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19530-418716/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19530-418716/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19530-418716/.minikube/certs/ca-key.pem org=jenkins.addons-505336 san=[127.0.0.1 192.168.49.2 addons-505336 localhost minikube]
I0829 18:56:16.986078 426808 provision.go:177] copyRemoteCerts
I0829 18:56:16.986138 426808 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0829 18:56:16.986171 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:17.004188 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:17.090811 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0829 18:56:17.111503 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0829 18:56:17.131929 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0829 18:56:17.152258 426808 provision.go:87] duration metric: took 289.525952ms to configureAuth
I0829 18:56:17.152282 426808 ubuntu.go:193] setting minikube options for container-runtime
I0829 18:56:17.152454 426808 config.go:182] Loaded profile config "addons-505336": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0829 18:56:17.152639 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:17.168078 426808 main.go:141] libmachine: Using SSH client type: native
I0829 18:56:17.168300 426808 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32803 <nil> <nil>}
I0829 18:56:17.168314 426808 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0829 18:56:17.290803 426808 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0829 18:56:17.290829 426808 ubuntu.go:71] root file system type: overlay
I0829 18:56:17.290938 426808 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0829 18:56:17.290991 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:17.307075 426808 main.go:141] libmachine: Using SSH client type: native
I0829 18:56:17.307243 426808 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32803 <nil> <nil>}
I0829 18:56:17.307304 426808 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0829 18:56:17.433063 426808 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0829 18:56:17.433135 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:17.448758 426808 main.go:141] libmachine: Using SSH client type: native
I0829 18:56:17.448927 426808 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32803 <nil> <nil>}
I0829 18:56:17.448943 426808 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0829 18:56:18.109112 426808 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-08-27 14:13:43.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-08-29 18:56:17.430331773 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0829 18:56:18.109138 426808 machine.go:96] duration metric: took 4.662053348s to provisionDockerMachine
I0829 18:56:18.109149 426808 client.go:171] duration metric: took 16.843251987s to LocalClient.Create
I0829 18:56:18.109168 426808 start.go:167] duration metric: took 16.84331394s to libmachine.API.Create "addons-505336"
I0829 18:56:18.109178 426808 start.go:293] postStartSetup for "addons-505336" (driver="docker")
I0829 18:56:18.109196 426808 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0829 18:56:18.109251 426808 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0829 18:56:18.109296 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:18.125308 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:18.215371 426808 ssh_runner.go:195] Run: cat /etc/os-release
I0829 18:56:18.218369 426808 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0829 18:56:18.218403 426808 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0829 18:56:18.218411 426808 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0829 18:56:18.218418 426808 info.go:137] Remote host: Ubuntu 22.04.4 LTS
I0829 18:56:18.218428 426808 filesync.go:126] Scanning /home/jenkins/minikube-integration/19530-418716/.minikube/addons for local assets ...
I0829 18:56:18.218484 426808 filesync.go:126] Scanning /home/jenkins/minikube-integration/19530-418716/.minikube/files for local assets ...
I0829 18:56:18.218509 426808 start.go:296] duration metric: took 109.320975ms for postStartSetup
I0829 18:56:18.218882 426808 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-505336
I0829 18:56:18.234288 426808 profile.go:143] Saving config to /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/config.json ...
I0829 18:56:18.234535 426808 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0829 18:56:18.234590 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:18.249930 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:18.339499 426808 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0829 18:56:18.343611 426808 start.go:128] duration metric: took 17.079841903s to createHost
I0829 18:56:18.343635 426808 start.go:83] releasing machines lock for "addons-505336", held for 17.079975128s
I0829 18:56:18.343708 426808 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-505336
I0829 18:56:18.358635 426808 ssh_runner.go:195] Run: cat /version.json
I0829 18:56:18.358679 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:18.358714 426808 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0829 18:56:18.358790 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:18.373806 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:18.376197 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:18.527058 426808 ssh_runner.go:195] Run: systemctl --version
I0829 18:56:18.531146 426808 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0829 18:56:18.534983 426808 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0829 18:56:18.556181 426808 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0829 18:56:18.556260 426808 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0829 18:56:18.580072 426808 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0829 18:56:18.580099 426808 start.go:495] detecting cgroup driver to use...
I0829 18:56:18.580132 426808 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0829 18:56:18.580242 426808 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0829 18:56:18.594135 426808 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0829 18:56:18.602652 426808 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0829 18:56:18.611000 426808 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0829 18:56:18.611047 426808 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0829 18:56:18.619188 426808 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0829 18:56:18.627298 426808 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0829 18:56:18.635103 426808 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0829 18:56:18.643536 426808 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0829 18:56:18.651134 426808 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0829 18:56:18.659093 426808 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0829 18:56:18.667374 426808 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0829 18:56:18.675638 426808 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0829 18:56:18.682531 426808 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0829 18:56:18.689606 426808 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0829 18:56:18.764743 426808 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0829 18:56:18.856717 426808 start.go:495] detecting cgroup driver to use...
I0829 18:56:18.856767 426808 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0829 18:56:18.856816 426808 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0829 18:56:18.867702 426808 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0829 18:56:18.867774 426808 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0829 18:56:18.878658 426808 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0829 18:56:18.893952 426808 ssh_runner.go:195] Run: which cri-dockerd
I0829 18:56:18.897109 426808 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0829 18:56:18.905495 426808 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0829 18:56:18.921675 426808 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0829 18:56:19.014652 426808 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0829 18:56:19.113584 426808 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0829 18:56:19.113724 426808 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0829 18:56:19.129567 426808 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0829 18:56:19.207284 426808 ssh_runner.go:195] Run: sudo systemctl restart docker
I0829 18:56:19.457471 426808 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0829 18:56:19.468165 426808 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0829 18:56:19.478588 426808 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0829 18:56:19.552351 426808 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0829 18:56:19.624515 426808 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0829 18:56:19.695986 426808 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0829 18:56:19.708259 426808 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0829 18:56:19.717891 426808 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0829 18:56:19.788321 426808 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0829 18:56:19.846455 426808 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0829 18:56:19.846559 426808 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0829 18:56:19.849981 426808 start.go:563] Will wait 60s for crictl version
I0829 18:56:19.850036 426808 ssh_runner.go:195] Run: which crictl
I0829 18:56:19.853141 426808 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0829 18:56:19.883049 426808 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.2.0
RuntimeApiVersion: v1
I0829 18:56:19.883120 426808 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0829 18:56:19.904937 426808 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0829 18:56:19.929612 426808 out.go:235] * Preparing Kubernetes v1.31.0 on Docker 27.2.0 ...
I0829 18:56:19.929680 426808 cli_runner.go:164] Run: docker network inspect addons-505336 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0829 18:56:19.944677 426808 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0829 18:56:19.947966 426808 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0829 18:56:19.957576 426808 kubeadm.go:883] updating cluster {Name:addons-505336 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-505336 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0829 18:56:19.957706 426808 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0829 18:56:19.957762 426808 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0829 18:56:19.977077 426808 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-scheduler:v1.31.0
registry.k8s.io/kube-apiserver:v1.31.0
registry.k8s.io/kube-controller-manager:v1.31.0
registry.k8s.io/kube-proxy:v1.31.0
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
registry.k8s.io/coredns/coredns:v1.11.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0829 18:56:19.977101 426808 docker.go:615] Images already preloaded, skipping extraction
I0829 18:56:19.977174 426808 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0829 18:56:19.994408 426808 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-controller-manager:v1.31.0
registry.k8s.io/kube-scheduler:v1.31.0
registry.k8s.io/kube-apiserver:v1.31.0
registry.k8s.io/kube-proxy:v1.31.0
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
registry.k8s.io/coredns/coredns:v1.11.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0829 18:56:19.994434 426808 cache_images.go:84] Images are preloaded, skipping loading
I0829 18:56:19.994455 426808 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.0 docker true true} ...
I0829 18:56:19.994605 426808 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.0/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-505336 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.0 ClusterName:addons-505336 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0829 18:56:19.994671 426808 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0829 18:56:20.037716 426808 cni.go:84] Creating CNI manager for ""
I0829 18:56:20.037748 426808 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0829 18:56:20.037767 426808 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0829 18:56:20.037800 426808 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.0 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-505336 NodeName:addons-505336 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0829 18:56:20.037966 426808 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-505336"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.0
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0829 18:56:20.038041 426808 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.0
I0829 18:56:20.045924 426808 binaries.go:44] Found k8s binaries, skipping transfer
I0829 18:56:20.045993 426808 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0829 18:56:20.053335 426808 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0829 18:56:20.068847 426808 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0829 18:56:20.084159 426808 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0829 18:56:20.099126 426808 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0829 18:56:20.102032 426808 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0829 18:56:20.111310 426808 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0829 18:56:20.192718 426808 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0829 18:56:20.204662 426808 certs.go:68] Setting up /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336 for IP: 192.168.49.2
I0829 18:56:20.204681 426808 certs.go:194] generating shared ca certs ...
I0829 18:56:20.204695 426808 certs.go:226] acquiring lock for ca certs: {Name:mka4e5df4d0f5dd863b35d0a189a931cab4268f0 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.204810 426808 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19530-418716/.minikube/ca.key
I0829 18:56:20.285035 426808 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19530-418716/.minikube/ca.crt ...
I0829 18:56:20.285062 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/ca.crt: {Name:mkdae7ace7bad9837a7e9117447db1af6f4b9c12 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.285221 426808 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19530-418716/.minikube/ca.key ...
I0829 18:56:20.285232 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/ca.key: {Name:mk653c9ed4efa28f38e9fb9d2fd0a214ec7239b7 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.285299 426808 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19530-418716/.minikube/proxy-client-ca.key
I0829 18:56:20.487866 426808 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19530-418716/.minikube/proxy-client-ca.crt ...
I0829 18:56:20.487901 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/proxy-client-ca.crt: {Name:mk3ea79acdff78220a070a6f46c75ff34ede82f0 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.488091 426808 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19530-418716/.minikube/proxy-client-ca.key ...
I0829 18:56:20.488112 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/proxy-client-ca.key: {Name:mke177191b156f83a8ba65e0fb9da46a36407c4b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.488193 426808 certs.go:256] generating profile certs ...
I0829 18:56:20.488264 426808 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/client.key
I0829 18:56:20.488280 426808 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/client.crt with IP's: []
I0829 18:56:20.602812 426808 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/client.crt ...
I0829 18:56:20.602838 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/client.crt: {Name:mk62766a01f914f2f1c3fd2f9596a5754163f207 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.602992 426808 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/client.key ...
I0829 18:56:20.603004 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/client.key: {Name:mk9b01a96eea1dafc4e09197e12633e073bde04b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.603073 426808 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.key.5b3d1d2d
I0829 18:56:20.603090 426808 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.crt.5b3d1d2d with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0829 18:56:20.709507 426808 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.crt.5b3d1d2d ...
I0829 18:56:20.709543 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.crt.5b3d1d2d: {Name:mk5807cc00835d78daa74d0205693bc70130675a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.709710 426808 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.key.5b3d1d2d ...
I0829 18:56:20.709725 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.key.5b3d1d2d: {Name:mk7be2718e8264d68c99ecad8d9bc4db42e631fc Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.709791 426808 certs.go:381] copying /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.crt.5b3d1d2d -> /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.crt
I0829 18:56:20.709858 426808 certs.go:385] copying /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.key.5b3d1d2d -> /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.key
I0829 18:56:20.709904 426808 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/proxy-client.key
I0829 18:56:20.709920 426808 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/proxy-client.crt with IP's: []
I0829 18:56:20.888091 426808 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/proxy-client.crt ...
I0829 18:56:20.888119 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/proxy-client.crt: {Name:mk2e01e96dbf706a41969e36b7af34b93c73e217 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.888267 426808 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/proxy-client.key ...
I0829 18:56:20.888280 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/proxy-client.key: {Name:mk2a869b5b77d7a6c475622ff109e619d913b029 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:20.888440 426808 certs.go:484] found cert: /home/jenkins/minikube-integration/19530-418716/.minikube/certs/ca-key.pem (1679 bytes)
I0829 18:56:20.888474 426808 certs.go:484] found cert: /home/jenkins/minikube-integration/19530-418716/.minikube/certs/ca.pem (1078 bytes)
I0829 18:56:20.888498 426808 certs.go:484] found cert: /home/jenkins/minikube-integration/19530-418716/.minikube/certs/cert.pem (1123 bytes)
I0829 18:56:20.888532 426808 certs.go:484] found cert: /home/jenkins/minikube-integration/19530-418716/.minikube/certs/key.pem (1675 bytes)
I0829 18:56:20.889204 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0829 18:56:20.910475 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0829 18:56:20.930979 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0829 18:56:20.951306 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I0829 18:56:20.971589 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0829 18:56:20.991477 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0829 18:56:21.012003 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0829 18:56:21.033235 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/profiles/addons-505336/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0829 18:56:21.054411 426808 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19530-418716/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0829 18:56:21.074680 426808 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0829 18:56:21.089925 426808 ssh_runner.go:195] Run: openssl version
I0829 18:56:21.094711 426808 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0829 18:56:21.103053 426808 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0829 18:56:21.106062 426808 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Aug 29 18:56 /usr/share/ca-certificates/minikubeCA.pem
I0829 18:56:21.106119 426808 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0829 18:56:21.112400 426808 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0829 18:56:21.120578 426808 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0829 18:56:21.123727 426808 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0829 18:56:21.123769 426808 kubeadm.go:392] StartCluster: {Name:addons-505336 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1724862063-19530@sha256:fd0f41868bf20a720502cce04c5201bfb064f3c267161af6fd5265d69c85c9f0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-505336 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0829 18:56:21.123871 426808 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0829 18:56:21.140346 426808 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0829 18:56:21.148140 426808 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0829 18:56:21.155884 426808 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0829 18:56:21.155927 426808 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0829 18:56:21.163186 426808 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0829 18:56:21.163202 426808 kubeadm.go:157] found existing configuration files:
I0829 18:56:21.163235 426808 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0829 18:56:21.170419 426808 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0829 18:56:21.170459 426808 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0829 18:56:21.177906 426808 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0829 18:56:21.185113 426808 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0829 18:56:21.185158 426808 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0829 18:56:21.192714 426808 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0829 18:56:21.199844 426808 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0829 18:56:21.199893 426808 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0829 18:56:21.207115 426808 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0829 18:56:21.214181 426808 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0829 18:56:21.214222 426808 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0829 18:56:21.221214 426808 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.0:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0829 18:56:21.252722 426808 kubeadm.go:310] W0829 18:56:21.252027 1922 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0829 18:56:21.253093 426808 kubeadm.go:310] W0829 18:56:21.252630 1922 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0829 18:56:21.272841 426808 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1067-gcp\n", err: exit status 1
I0829 18:56:21.320161 426808 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0829 18:56:30.174036 426808 kubeadm.go:310] [init] Using Kubernetes version: v1.31.0
I0829 18:56:30.174094 426808 kubeadm.go:310] [preflight] Running pre-flight checks
I0829 18:56:30.174174 426808 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0829 18:56:30.174242 426808 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1067-gcp[0m
I0829 18:56:30.174282 426808 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0829 18:56:30.174377 426808 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0829 18:56:30.174463 426808 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0829 18:56:30.174551 426808 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0829 18:56:30.174622 426808 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0829 18:56:30.174698 426808 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0829 18:56:30.174768 426808 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0829 18:56:30.174848 426808 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0829 18:56:30.174897 426808 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0829 18:56:30.174947 426808 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0829 18:56:30.175047 426808 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0829 18:56:30.175185 426808 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0829 18:56:30.175312 426808 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0829 18:56:30.175406 426808 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0829 18:56:30.176894 426808 out.go:235] - Generating certificates and keys ...
I0829 18:56:30.176995 426808 kubeadm.go:310] [certs] Using existing ca certificate authority
I0829 18:56:30.177090 426808 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0829 18:56:30.177154 426808 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0829 18:56:30.177234 426808 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0829 18:56:30.177333 426808 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0829 18:56:30.177392 426808 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0829 18:56:30.177438 426808 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0829 18:56:30.177586 426808 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-505336 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0829 18:56:30.177672 426808 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0829 18:56:30.177792 426808 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-505336 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0829 18:56:30.177864 426808 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0829 18:56:30.177932 426808 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0829 18:56:30.177990 426808 kubeadm.go:310] [certs] Generating "sa" key and public key
I0829 18:56:30.178071 426808 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0829 18:56:30.178124 426808 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0829 18:56:30.178172 426808 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0829 18:56:30.178216 426808 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0829 18:56:30.178269 426808 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0829 18:56:30.178324 426808 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0829 18:56:30.178391 426808 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0829 18:56:30.178471 426808 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0829 18:56:30.179556 426808 out.go:235] - Booting up control plane ...
I0829 18:56:30.179654 426808 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0829 18:56:30.179745 426808 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0829 18:56:30.179834 426808 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0829 18:56:30.179959 426808 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0829 18:56:30.180075 426808 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0829 18:56:30.180113 426808 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0829 18:56:30.180253 426808 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0829 18:56:30.180340 426808 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0829 18:56:30.180390 426808 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 501.650816ms
I0829 18:56:30.180483 426808 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0829 18:56:30.180580 426808 kubeadm.go:310] [api-check] The API server is healthy after 4.501290427s
I0829 18:56:30.180744 426808 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0829 18:56:30.180917 426808 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0829 18:56:30.181000 426808 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0829 18:56:30.181227 426808 kubeadm.go:310] [mark-control-plane] Marking the node addons-505336 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0829 18:56:30.181327 426808 kubeadm.go:310] [bootstrap-token] Using token: mazegd.fx0lflllcl8q5igp
I0829 18:56:30.182583 426808 out.go:235] - Configuring RBAC rules ...
I0829 18:56:30.182706 426808 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0829 18:56:30.182852 426808 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0829 18:56:30.183056 426808 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0829 18:56:30.183243 426808 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0829 18:56:30.183366 426808 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0829 18:56:30.183444 426808 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0829 18:56:30.183546 426808 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0829 18:56:30.183587 426808 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0829 18:56:30.183626 426808 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0829 18:56:30.183631 426808 kubeadm.go:310]
I0829 18:56:30.183683 426808 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0829 18:56:30.183689 426808 kubeadm.go:310]
I0829 18:56:30.183753 426808 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0829 18:56:30.183759 426808 kubeadm.go:310]
I0829 18:56:30.183785 426808 kubeadm.go:310] mkdir -p $HOME/.kube
I0829 18:56:30.183836 426808 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0829 18:56:30.183878 426808 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0829 18:56:30.183885 426808 kubeadm.go:310]
I0829 18:56:30.183927 426808 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0829 18:56:30.183936 426808 kubeadm.go:310]
I0829 18:56:30.183977 426808 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0829 18:56:30.183983 426808 kubeadm.go:310]
I0829 18:56:30.184028 426808 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0829 18:56:30.184089 426808 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0829 18:56:30.184148 426808 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0829 18:56:30.184154 426808 kubeadm.go:310]
I0829 18:56:30.184220 426808 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0829 18:56:30.184284 426808 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0829 18:56:30.184290 426808 kubeadm.go:310]
I0829 18:56:30.184382 426808 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token mazegd.fx0lflllcl8q5igp \
I0829 18:56:30.184465 426808 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:e9bfd7afbd259a94815ec69a6713b021add294d72979ae362a8e14d4b4c83f5e \
I0829 18:56:30.184490 426808 kubeadm.go:310] --control-plane
I0829 18:56:30.184496 426808 kubeadm.go:310]
I0829 18:56:30.184572 426808 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0829 18:56:30.184578 426808 kubeadm.go:310]
I0829 18:56:30.184650 426808 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token mazegd.fx0lflllcl8q5igp \
I0829 18:56:30.184776 426808 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:e9bfd7afbd259a94815ec69a6713b021add294d72979ae362a8e14d4b4c83f5e
I0829 18:56:30.184791 426808 cni.go:84] Creating CNI manager for ""
I0829 18:56:30.184803 426808 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0829 18:56:30.186008 426808 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0829 18:56:30.187135 426808 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0829 18:56:30.195381 426808 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0829 18:56:30.210693 426808 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0829 18:56:30.210747 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:30.210798 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-505336 minikube.k8s.io/updated_at=2024_08_29T18_56_30_0700 minikube.k8s.io/version=v1.33.1 minikube.k8s.io/commit=5512bd76519cf55fa04aeca1cd01a1369e298033 minikube.k8s.io/name=addons-505336 minikube.k8s.io/primary=true
I0829 18:56:30.271729 426808 ops.go:34] apiserver oom_adj: -16
I0829 18:56:30.271754 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:30.771818 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:31.271826 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:31.772347 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:32.272029 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:32.772814 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:33.272599 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:33.772465 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:34.272051 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:34.772813 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:35.272776 426808 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0829 18:56:35.332994 426808 kubeadm.go:1113] duration metric: took 5.122290558s to wait for elevateKubeSystemPrivileges
I0829 18:56:35.333036 426808 kubeadm.go:394] duration metric: took 14.209271406s to StartCluster
I0829 18:56:35.333062 426808 settings.go:142] acquiring lock: {Name:mk8d18e194f03e62292d34a90cdfbd838fafb153 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:35.333216 426808 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19530-418716/kubeconfig
I0829 18:56:35.333607 426808 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19530-418716/kubeconfig: {Name:mk28ee8c6d7b073bfc329febc7f1844f1691b19d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0829 18:56:35.333842 426808 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0829 18:56:35.333854 426808 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0829 18:56:35.333913 426808 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0829 18:56:35.334007 426808 addons.go:69] Setting yakd=true in profile "addons-505336"
I0829 18:56:35.334022 426808 addons.go:69] Setting cloud-spanner=true in profile "addons-505336"
I0829 18:56:35.334043 426808 addons.go:234] Setting addon yakd=true in "addons-505336"
I0829 18:56:35.334057 426808 config.go:182] Loaded profile config "addons-505336": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0829 18:56:35.334055 426808 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-505336"
I0829 18:56:35.334064 426808 addons.go:69] Setting metrics-server=true in profile "addons-505336"
I0829 18:56:35.334078 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.334083 426808 addons.go:69] Setting registry=true in profile "addons-505336"
I0829 18:56:35.334092 426808 addons.go:69] Setting volcano=true in profile "addons-505336"
I0829 18:56:35.334094 426808 addons.go:234] Setting addon metrics-server=true in "addons-505336"
I0829 18:56:35.334103 426808 addons.go:234] Setting addon registry=true in "addons-505336"
I0829 18:56:35.334111 426808 addons.go:234] Setting addon volcano=true in "addons-505336"
I0829 18:56:35.334124 426808 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-505336"
I0829 18:56:35.334127 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.334131 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.334133 426808 addons.go:69] Setting gcp-auth=true in profile "addons-505336"
I0829 18:56:35.334149 426808 mustload.go:65] Loading cluster: addons-505336
I0829 18:56:35.334150 426808 addons.go:69] Setting volumesnapshots=true in profile "addons-505336"
I0829 18:56:35.334154 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.334171 426808 addons.go:234] Setting addon volumesnapshots=true in "addons-505336"
I0829 18:56:35.334195 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.334335 426808 config.go:182] Loaded profile config "addons-505336": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0829 18:56:35.334361 426808 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-505336"
I0829 18:56:35.334411 426808 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-505336"
I0829 18:56:35.334571 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.334625 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.334636 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.334642 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.334650 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.334652 426808 addons.go:69] Setting ingress=true in profile "addons-505336"
I0829 18:56:35.334677 426808 addons.go:234] Setting addon ingress=true in "addons-505336"
I0829 18:56:35.334690 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.334709 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.334777 426808 addons.go:69] Setting default-storageclass=true in profile "addons-505336"
I0829 18:56:35.334850 426808 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-505336"
I0829 18:56:35.335135 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.334068 426808 addons.go:234] Setting addon cloud-spanner=true in "addons-505336"
I0829 18:56:35.335188 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.335242 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.334126 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.335140 426808 addons.go:69] Setting ingress-dns=true in profile "addons-505336"
I0829 18:56:35.335369 426808 addons.go:234] Setting addon ingress-dns=true in "addons-505336"
I0829 18:56:35.335403 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.334642 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.335152 426808 addons.go:69] Setting helm-tiller=true in profile "addons-505336"
I0829 18:56:35.335771 426808 addons.go:234] Setting addon helm-tiller=true in "addons-505336"
I0829 18:56:35.335797 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.335822 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.334076 426808 addons.go:69] Setting storage-provisioner=true in profile "addons-505336"
I0829 18:56:35.335959 426808 addons.go:234] Setting addon storage-provisioner=true in "addons-505336"
I0829 18:56:35.335982 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.336562 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.334080 426808 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-505336"
I0829 18:56:35.336659 426808 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-505336"
I0829 18:56:35.336689 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.335160 426808 addons.go:69] Setting inspektor-gadget=true in profile "addons-505336"
I0829 18:56:35.336716 426808 addons.go:234] Setting addon inspektor-gadget=true in "addons-505336"
I0829 18:56:35.336753 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.337102 426808 out.go:177] * Verifying Kubernetes components...
I0829 18:56:35.337170 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.337130 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.338372 426808 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0829 18:56:35.366673 426808 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-505336"
I0829 18:56:35.366728 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.367411 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.370831 426808 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0829 18:56:35.370934 426808 out.go:177] - Using image docker.io/registry:2.8.3
I0829 18:56:35.372000 426808 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0829 18:56:35.372019 426808 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0829 18:56:35.372073 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.372367 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.372801 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.373186 426808 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0829 18:56:35.373635 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.374559 426808 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0829 18:56:35.374578 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0829 18:56:35.374620 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.379028 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.382324 426808 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0829 18:56:35.383498 426808 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0829 18:56:35.384525 426808 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0829 18:56:35.386648 426808 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0829 18:56:35.386683 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0829 18:56:35.386739 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.386906 426808 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0829 18:56:35.386911 426808 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0829 18:56:35.388006 426808 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0829 18:56:35.388024 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0829 18:56:35.388077 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.389339 426808 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0829 18:56:35.390302 426808 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0829 18:56:35.391256 426808 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0829 18:56:35.392178 426808 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.31.0
I0829 18:56:35.392267 426808 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0829 18:56:35.393998 426808 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0829 18:56:35.394013 426808 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0829 18:56:35.394069 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.398813 426808 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0829 18:56:35.401987 426808 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0829 18:56:35.402096 426808 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0829 18:56:35.403785 426808 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0829 18:56:35.403804 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0829 18:56:35.403856 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.405394 426808 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0829 18:56:35.406443 426808 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0829 18:56:35.406462 426808 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0829 18:56:35.406515 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.429902 426808 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0829 18:56:35.430935 426808 out.go:177] - Using image docker.io/busybox:stable
I0829 18:56:35.432003 426808 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0829 18:56:35.432020 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0829 18:56:35.432074 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.434592 426808 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0829 18:56:35.435733 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.437368 426808 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0829 18:56:35.438759 426808 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0829 18:56:35.440864 426808 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0829 18:56:35.440885 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0829 18:56:35.440938 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.444419 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.448881 426808 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0829 18:56:35.448918 426808 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0829 18:56:35.449324 426808 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.1
I0829 18:56:35.450393 426808 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0829 18:56:35.450411 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0829 18:56:35.450464 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.450869 426808 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0829 18:56:35.450888 426808 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0829 18:56:35.450942 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.451093 426808 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0829 18:56:35.451104 426808 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0829 18:56:35.451144 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.454141 426808 addons.go:234] Setting addon default-storageclass=true in "addons-505336"
I0829 18:56:35.454180 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:35.454744 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:35.458186 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.458949 426808 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0829 18:56:35.460035 426808 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0829 18:56:35.462929 426808 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0829 18:56:35.462949 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0829 18:56:35.463005 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.463640 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.463700 426808 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0829 18:56:35.463711 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0829 18:56:35.463753 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:35.469266 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.484165 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.484641 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.509616 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.511539 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.512060 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.512075 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.514659 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.516044 426808 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0829 18:56:35.516063 426808 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0829 18:56:35.516116 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
W0829 18:56:35.516329 426808 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0829 18:56:35.516362 426808 retry.go:31] will retry after 328.783396ms: ssh: handshake failed: EOF
I0829 18:56:35.517173 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.524846 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.535321 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:35.681881 426808 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0829 18:56:35.682023 426808 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0829 18:56:35.802328 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0829 18:56:35.897615 426808 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0829 18:56:35.897716 426808 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0829 18:56:36.091778 426808 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0829 18:56:36.091811 426808 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0829 18:56:36.183122 426808 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0829 18:56:36.183157 426808 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0829 18:56:36.191400 426808 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0829 18:56:36.191425 426808 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0829 18:56:36.195855 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0829 18:56:36.201633 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0829 18:56:36.276795 426808 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0829 18:56:36.276887 426808 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0829 18:56:36.279960 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0829 18:56:36.280789 426808 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0829 18:56:36.280844 426808 ssh_runner.go:362] scp helm-tiller/helm-tiller-rbac.yaml --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0829 18:56:36.294225 426808 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0829 18:56:36.294304 426808 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0829 18:56:36.294905 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0829 18:56:36.296437 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0829 18:56:36.377509 426808 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0829 18:56:36.377541 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0829 18:56:36.377859 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0829 18:56:36.380438 426808 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0829 18:56:36.380470 426808 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0829 18:56:36.481851 426808 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0829 18:56:36.481933 426808 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0829 18:56:36.587189 426808 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0829 18:56:36.587266 426808 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0829 18:56:36.594523 426808 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0829 18:56:36.594551 426808 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0829 18:56:36.684995 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0829 18:56:36.688956 426808 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0829 18:56:36.689054 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0829 18:56:36.695605 426808 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0829 18:56:36.695674 426808 ssh_runner.go:362] scp helm-tiller/helm-tiller-svc.yaml --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0829 18:56:36.696903 426808 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0829 18:56:36.696992 426808 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0829 18:56:36.781703 426808 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0829 18:56:36.781792 426808 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0829 18:56:36.977960 426808 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0829 18:56:36.978052 426808 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0829 18:56:36.994014 426808 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0829 18:56:36.994099 426808 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0829 18:56:37.178026 426808 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0829 18:56:37.178099 426808 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0829 18:56:37.195123 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0829 18:56:37.294310 426808 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0829 18:56:37.294392 426808 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0829 18:56:37.296863 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0829 18:56:37.479354 426808 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.797428008s)
I0829 18:56:37.479562 426808 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0829 18:56:37.479470 426808 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.797418293s)
I0829 18:56:37.480795 426808 node_ready.go:35] waiting up to 6m0s for node "addons-505336" to be "Ready" ...
I0829 18:56:37.485158 426808 node_ready.go:49] node "addons-505336" has status "Ready":"True"
I0829 18:56:37.485244 426808 node_ready.go:38] duration metric: took 4.371801ms for node "addons-505336" to be "Ready" ...
I0829 18:56:37.485279 426808 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0829 18:56:37.493580 426808 pod_ready.go:79] waiting up to 6m0s for pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace to be "Ready" ...
I0829 18:56:37.584164 426808 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0829 18:56:37.584194 426808 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0829 18:56:37.679167 426808 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0829 18:56:37.679194 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0829 18:56:37.777704 426808 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0829 18:56:37.777734 426808 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0829 18:56:37.880553 426808 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0829 18:56:37.880644 426808 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0829 18:56:37.881638 426808 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0829 18:56:37.881700 426808 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0829 18:56:37.994980 426808 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-505336" context rescaled to 1 replicas
I0829 18:56:38.199017 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0829 18:56:38.277601 426808 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0829 18:56:38.277700 426808 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0829 18:56:38.281835 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0829 18:56:38.480630 426808 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0829 18:56:38.480733 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0829 18:56:38.485060 426808 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0829 18:56:38.485138 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0829 18:56:38.776468 426808 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0829 18:56:38.776503 426808 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0829 18:56:38.980619 426808 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0829 18:56:38.980716 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0829 18:56:39.180002 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0829 18:56:39.282607 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (3.086713141s)
I0829 18:56:39.282739 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (3.081074131s)
I0829 18:56:39.282920 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (3.480556806s)
I0829 18:56:39.583927 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:56:39.686478 426808 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0829 18:56:39.686571 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0829 18:56:39.798983 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0829 18:56:40.292223 426808 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0829 18:56:40.292317 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0829 18:56:41.080835 426808 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0829 18:56:41.080918 426808 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0829 18:56:41.592340 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0829 18:56:42.077672 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:56:42.388913 426808 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0829 18:56:42.389062 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:42.411321 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:43.583533 426808 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0829 18:56:43.794757 426808 addons.go:234] Setting addon gcp-auth=true in "addons-505336"
I0829 18:56:43.794841 426808 host.go:66] Checking if "addons-505336" exists ...
I0829 18:56:43.795342 426808 cli_runner.go:164] Run: docker container inspect addons-505336 --format={{.State.Status}}
I0829 18:56:43.813591 426808 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0829 18:56:43.813642 426808 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-505336
I0829 18:56:43.828758 426808 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32803 SSHKeyPath:/home/jenkins/minikube-integration/19530-418716/.minikube/machines/addons-505336/id_rsa Username:docker}
I0829 18:56:44.083443 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:56:46.498959 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:56:47.887540 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.607499979s)
I0829 18:56:47.887751 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (11.592788253s)
I0829 18:56:47.887791 426808 addons.go:475] Verifying addon ingress=true in "addons-505336"
I0829 18:56:47.888042 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (11.591532083s)
I0829 18:56:47.888288 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (10.693054549s)
I0829 18:56:47.888355 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (10.591418267s)
I0829 18:56:47.888369 426808 addons.go:475] Verifying addon registry=true in "addons-505336"
I0829 18:56:47.888164 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (11.510271339s)
I0829 18:56:47.888214 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (11.203195522s)
I0829 18:56:47.888573 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (9.689515456s)
I0829 18:56:47.888692 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (9.6067637s)
I0829 18:56:47.888727 426808 addons.go:475] Verifying addon metrics-server=true in "addons-505336"
I0829 18:56:47.888828 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (8.708739548s)
W0829 18:56:47.888867 426808 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0829 18:56:47.888892 426808 retry.go:31] will retry after 243.453168ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0829 18:56:47.889021 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (8.08993811s)
I0829 18:56:47.889557 426808 out.go:177] * Verifying registry addon...
I0829 18:56:47.890460 426808 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-505336 service yakd-dashboard -n yakd-dashboard
I0829 18:56:47.890560 426808 out.go:177] * Verifying ingress addon...
I0829 18:56:47.892602 426808 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0829 18:56:47.893804 426808 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0829 18:56:47.898437 426808 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0829 18:56:47.898458 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:47.899780 426808 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0829 18:56:47.899802 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
W0829 18:56:47.980471 426808 out.go:270] ! Enabling 'storage-provisioner-rancher' returned an error: running callbacks: [Error making local-path the default storage class: Error while marking storage class local-path as default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
I0829 18:56:48.133319 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0829 18:56:48.398857 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:48.476098 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:48.588637 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:56:48.901001 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:48.901892 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:49.400521 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:49.400869 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:49.575965 426808 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.762339839s)
I0829 18:56:49.576177 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (7.983463959s)
I0829 18:56:49.576264 426808 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-505336"
I0829 18:56:49.577562 426808 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0829 18:56:49.578680 426808 out.go:177] * Verifying csi-hostpath-driver addon...
I0829 18:56:49.583749 426808 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0829 18:56:49.584846 426808 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0829 18:56:49.585520 426808 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0829 18:56:49.585550 426808 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0829 18:56:49.591140 426808 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0829 18:56:49.591206 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:49.685027 426808 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0829 18:56:49.685054 426808 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0829 18:56:49.708214 426808 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0829 18:56:49.708244 426808 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0829 18:56:49.793075 426808 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0829 18:56:49.898356 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:49.899061 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:50.091393 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:50.397459 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:50.398594 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:50.590749 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:50.790701 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.657284517s)
I0829 18:56:50.896334 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:50.898257 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:51.000053 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:56:51.089146 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:51.104637 426808 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.311516707s)
I0829 18:56:51.106502 426808 addons.go:475] Verifying addon gcp-auth=true in "addons-505336"
I0829 18:56:51.108092 426808 out.go:177] * Verifying gcp-auth addon...
I0829 18:56:51.110152 426808 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0829 18:56:51.187977 426808 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0829 18:56:51.396594 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:51.397609 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:51.590110 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:51.896825 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:51.897950 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:52.089969 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:52.397058 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:52.397784 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:52.590007 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:52.896777 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:52.897554 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:53.089799 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:53.396768 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:53.398020 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:53.499100 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:56:53.589725 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:53.896900 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:53.897957 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:54.089655 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:54.396261 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:54.397422 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:54.588823 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:54.898007 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:54.898354 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:55.090652 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:55.397114 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:55.397572 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:55.499388 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:56:55.589233 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:55.896412 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:55.897369 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:56.089084 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:56.396454 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:56.397607 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:56.589497 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:56.896715 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:56.897272 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:57.089500 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:57.449551 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:57.450116 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:57.588589 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:57.896689 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:57.897706 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:57.999622 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:56:58.089044 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:58.396642 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:58.397199 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:58.590410 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:58.896721 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:58.897884 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:59.090145 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:59.396159 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:59.397944 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:56:59.590255 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:56:59.896168 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:56:59.898172 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:00.000308 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:57:00.090776 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:00.396121 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:00.398200 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:00.590030 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:00.898207 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:00.898602 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:01.090080 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:01.396073 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:01.397624 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:01.589669 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:01.896770 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:01.897336 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:02.090220 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:02.396668 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:02.398358 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:02.499610 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:57:02.589935 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:02.896500 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:02.898152 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:03.090153 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:03.396322 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:03.398029 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:03.589075 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:03.895893 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:03.897857 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:04.090489 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:04.397092 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:04.398044 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:04.501680 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:57:04.590582 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:04.896201 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:04.897389 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:05.089941 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:05.397440 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:05.398331 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:05.589727 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:05.897040 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:05.897901 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:06.089733 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:06.396286 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:06.398417 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:06.590070 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:06.896276 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:06.898162 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:06.999092 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:57:07.088755 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:07.396255 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:07.399197 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:07.589829 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:07.896980 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:07.897422 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:08.090101 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:08.397121 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:08.397653 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:08.589408 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:08.897262 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:08.897849 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:08.999742 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:57:09.090094 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:09.396857 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:09.397843 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:09.589274 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:09.897114 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:09.897768 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:10.089644 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:10.396666 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:10.397801 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:10.589385 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:10.896137 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:10.897205 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:11.000452 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:57:11.089351 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:11.396304 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0829 18:57:11.398016 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:11.589549 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:11.897400 426808 kapi.go:107] duration metric: took 24.004796963s to wait for kubernetes.io/minikube-addons=registry ...
I0829 18:57:11.897875 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:12.090003 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:12.456228 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:12.590527 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:12.903269 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:13.000977 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:57:13.089964 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:13.397957 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:13.590415 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:13.898191 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:14.089147 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:14.399103 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:14.590227 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:14.901536 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:15.002801 426808 pod_ready.go:103] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"False"
I0829 18:57:15.089915 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:15.398495 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:15.503433 426808 pod_ready.go:93] pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace has status "Ready":"True"
I0829 18:57:15.503462 426808 pod_ready.go:82] duration metric: took 38.009790923s for pod "coredns-6f6b679f8f-2c86p" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.503491 426808 pod_ready.go:79] waiting up to 6m0s for pod "coredns-6f6b679f8f-9ljnh" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.505070 426808 pod_ready.go:98] error getting pod "coredns-6f6b679f8f-9ljnh" in "kube-system" namespace (skipping!): pods "coredns-6f6b679f8f-9ljnh" not found
I0829 18:57:15.505095 426808 pod_ready.go:82] duration metric: took 1.595594ms for pod "coredns-6f6b679f8f-9ljnh" in "kube-system" namespace to be "Ready" ...
E0829 18:57:15.505107 426808 pod_ready.go:67] WaitExtra: waitPodCondition: error getting pod "coredns-6f6b679f8f-9ljnh" in "kube-system" namespace (skipping!): pods "coredns-6f6b679f8f-9ljnh" not found
I0829 18:57:15.505115 426808 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-505336" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.509431 426808 pod_ready.go:93] pod "etcd-addons-505336" in "kube-system" namespace has status "Ready":"True"
I0829 18:57:15.509451 426808 pod_ready.go:82] duration metric: took 4.327516ms for pod "etcd-addons-505336" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.509462 426808 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-505336" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.514278 426808 pod_ready.go:93] pod "kube-apiserver-addons-505336" in "kube-system" namespace has status "Ready":"True"
I0829 18:57:15.514298 426808 pod_ready.go:82] duration metric: took 4.828284ms for pod "kube-apiserver-addons-505336" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.514309 426808 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-505336" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.518115 426808 pod_ready.go:93] pod "kube-controller-manager-addons-505336" in "kube-system" namespace has status "Ready":"True"
I0829 18:57:15.518137 426808 pod_ready.go:82] duration metric: took 3.820399ms for pod "kube-controller-manager-addons-505336" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.518150 426808 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-kj5d4" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.589526 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:15.697320 426808 pod_ready.go:93] pod "kube-proxy-kj5d4" in "kube-system" namespace has status "Ready":"True"
I0829 18:57:15.697341 426808 pod_ready.go:82] duration metric: took 179.183733ms for pod "kube-proxy-kj5d4" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.697350 426808 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-505336" in "kube-system" namespace to be "Ready" ...
I0829 18:57:15.898213 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:16.090330 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:16.097781 426808 pod_ready.go:93] pod "kube-scheduler-addons-505336" in "kube-system" namespace has status "Ready":"True"
I0829 18:57:16.097809 426808 pod_ready.go:82] duration metric: took 400.447898ms for pod "kube-scheduler-addons-505336" in "kube-system" namespace to be "Ready" ...
I0829 18:57:16.097819 426808 pod_ready.go:39] duration metric: took 38.612494425s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0829 18:57:16.097842 426808 api_server.go:52] waiting for apiserver process to appear ...
I0829 18:57:16.097898 426808 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0829 18:57:16.116949 426808 api_server.go:72] duration metric: took 40.783056601s to wait for apiserver process to appear ...
I0829 18:57:16.116974 426808 api_server.go:88] waiting for apiserver healthz status ...
I0829 18:57:16.116994 426808 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0829 18:57:16.121123 426808 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0829 18:57:16.121968 426808 api_server.go:141] control plane version: v1.31.0
I0829 18:57:16.121993 426808 api_server.go:131] duration metric: took 5.009702ms to wait for apiserver health ...
I0829 18:57:16.122003 426808 system_pods.go:43] waiting for kube-system pods to appear ...
I0829 18:57:16.303461 426808 system_pods.go:59] 18 kube-system pods found
I0829 18:57:16.303501 426808 system_pods.go:61] "coredns-6f6b679f8f-2c86p" [f67d4ad9-1590-4d60-bc65-99d35dc95936] Running
I0829 18:57:16.303513 426808 system_pods.go:61] "csi-hostpath-attacher-0" [775fcb58-3d17-43c4-b475-7649fcd6e015] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0829 18:57:16.303520 426808 system_pods.go:61] "csi-hostpath-resizer-0" [16bfb94a-a8a7-4334-bcbe-7bc1dd42061d] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0829 18:57:16.303528 426808 system_pods.go:61] "csi-hostpathplugin-p6tdr" [f1ddb0db-9983-4573-832c-e588b9bad378] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0829 18:57:16.303534 426808 system_pods.go:61] "etcd-addons-505336" [fffeff5b-44c0-4db0-8734-7e95b2a2793f] Running
I0829 18:57:16.303540 426808 system_pods.go:61] "kube-apiserver-addons-505336" [a850d56d-bf42-4cb1-99ae-482755049de1] Running
I0829 18:57:16.303547 426808 system_pods.go:61] "kube-controller-manager-addons-505336" [a960f230-49c7-4378-ada6-812b94c4eb93] Running
I0829 18:57:16.303553 426808 system_pods.go:61] "kube-ingress-dns-minikube" [9b161863-95d0-4562-9c1a-9ebb464a71f7] Running
I0829 18:57:16.303558 426808 system_pods.go:61] "kube-proxy-kj5d4" [03308cc4-469b-465a-bb58-8585e006fb32] Running
I0829 18:57:16.303562 426808 system_pods.go:61] "kube-scheduler-addons-505336" [1bd1dd8f-1518-4b86-a726-ea86fd6a8b8b] Running
I0829 18:57:16.303565 426808 system_pods.go:61] "metrics-server-8988944d9-brsc8" [edb7fd7c-feae-493c-8e00-a0629ed0235b] Running
I0829 18:57:16.303568 426808 system_pods.go:61] "nvidia-device-plugin-daemonset-vjlkn" [56fd832b-ce49-4077-95a5-fc21a8abdc0d] Running
I0829 18:57:16.303572 426808 system_pods.go:61] "registry-6fb4cdfc84-7xs8r" [0bc8f454-eced-450f-ab5d-b961648307b9] Running
I0829 18:57:16.303579 426808 system_pods.go:61] "registry-proxy-wwxwv" [ef74c371-cbb9-4c81-91dd-dcbc748f81d0] Running
I0829 18:57:16.303588 426808 system_pods.go:61] "snapshot-controller-56fcc65765-5zcjx" [9fbc68ee-15f3-4427-a51a-61672bef2410] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0829 18:57:16.303600 426808 system_pods.go:61] "snapshot-controller-56fcc65765-qp5gp" [7b2a8697-1afe-40de-a372-01913a79e7bf] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0829 18:57:16.303609 426808 system_pods.go:61] "storage-provisioner" [617760c7-e1b6-4f22-bd1c-a7bece49bde3] Running
I0829 18:57:16.303615 426808 system_pods.go:61] "tiller-deploy-b48cc5f79-7x5zs" [32e00d4f-af90-4a61-9f18-048d53bb045d] Running
I0829 18:57:16.303623 426808 system_pods.go:74] duration metric: took 181.612484ms to wait for pod list to return data ...
I0829 18:57:16.303634 426808 default_sa.go:34] waiting for default service account to be created ...
I0829 18:57:16.398599 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:16.497695 426808 default_sa.go:45] found service account: "default"
I0829 18:57:16.497719 426808 default_sa.go:55] duration metric: took 194.077477ms for default service account to be created ...
I0829 18:57:16.497729 426808 system_pods.go:116] waiting for k8s-apps to be running ...
I0829 18:57:16.589686 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:16.704475 426808 system_pods.go:86] 18 kube-system pods found
I0829 18:57:16.704516 426808 system_pods.go:89] "coredns-6f6b679f8f-2c86p" [f67d4ad9-1590-4d60-bc65-99d35dc95936] Running
I0829 18:57:16.704531 426808 system_pods.go:89] "csi-hostpath-attacher-0" [775fcb58-3d17-43c4-b475-7649fcd6e015] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0829 18:57:16.704540 426808 system_pods.go:89] "csi-hostpath-resizer-0" [16bfb94a-a8a7-4334-bcbe-7bc1dd42061d] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0829 18:57:16.704551 426808 system_pods.go:89] "csi-hostpathplugin-p6tdr" [f1ddb0db-9983-4573-832c-e588b9bad378] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0829 18:57:16.704557 426808 system_pods.go:89] "etcd-addons-505336" [fffeff5b-44c0-4db0-8734-7e95b2a2793f] Running
I0829 18:57:16.704563 426808 system_pods.go:89] "kube-apiserver-addons-505336" [a850d56d-bf42-4cb1-99ae-482755049de1] Running
I0829 18:57:16.704570 426808 system_pods.go:89] "kube-controller-manager-addons-505336" [a960f230-49c7-4378-ada6-812b94c4eb93] Running
I0829 18:57:16.704581 426808 system_pods.go:89] "kube-ingress-dns-minikube" [9b161863-95d0-4562-9c1a-9ebb464a71f7] Running
I0829 18:57:16.704587 426808 system_pods.go:89] "kube-proxy-kj5d4" [03308cc4-469b-465a-bb58-8585e006fb32] Running
I0829 18:57:16.704595 426808 system_pods.go:89] "kube-scheduler-addons-505336" [1bd1dd8f-1518-4b86-a726-ea86fd6a8b8b] Running
I0829 18:57:16.704601 426808 system_pods.go:89] "metrics-server-8988944d9-brsc8" [edb7fd7c-feae-493c-8e00-a0629ed0235b] Running
I0829 18:57:16.704612 426808 system_pods.go:89] "nvidia-device-plugin-daemonset-vjlkn" [56fd832b-ce49-4077-95a5-fc21a8abdc0d] Running
I0829 18:57:16.704620 426808 system_pods.go:89] "registry-6fb4cdfc84-7xs8r" [0bc8f454-eced-450f-ab5d-b961648307b9] Running
I0829 18:57:16.704628 426808 system_pods.go:89] "registry-proxy-wwxwv" [ef74c371-cbb9-4c81-91dd-dcbc748f81d0] Running
I0829 18:57:16.704638 426808 system_pods.go:89] "snapshot-controller-56fcc65765-5zcjx" [9fbc68ee-15f3-4427-a51a-61672bef2410] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0829 18:57:16.704649 426808 system_pods.go:89] "snapshot-controller-56fcc65765-qp5gp" [7b2a8697-1afe-40de-a372-01913a79e7bf] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0829 18:57:16.704656 426808 system_pods.go:89] "storage-provisioner" [617760c7-e1b6-4f22-bd1c-a7bece49bde3] Running
I0829 18:57:16.704665 426808 system_pods.go:89] "tiller-deploy-b48cc5f79-7x5zs" [32e00d4f-af90-4a61-9f18-048d53bb045d] Running
I0829 18:57:16.704676 426808 system_pods.go:126] duration metric: took 206.938947ms to wait for k8s-apps to be running ...
I0829 18:57:16.704691 426808 system_svc.go:44] waiting for kubelet service to be running ....
I0829 18:57:16.704743 426808 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0829 18:57:16.718811 426808 system_svc.go:56] duration metric: took 14.091997ms WaitForService to wait for kubelet
I0829 18:57:16.718886 426808 kubeadm.go:582] duration metric: took 41.384996745s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0829 18:57:16.718936 426808 node_conditions.go:102] verifying NodePressure condition ...
I0829 18:57:16.897100 426808 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0829 18:57:16.897132 426808 node_conditions.go:123] node cpu capacity is 8
I0829 18:57:16.897145 426808 node_conditions.go:105] duration metric: took 178.203156ms to run NodePressure ...
I0829 18:57:16.897160 426808 start.go:241] waiting for startup goroutines ...
I0829 18:57:16.897804 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:17.090245 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:17.398329 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:17.590135 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:17.899518 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:18.089925 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:18.398215 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:18.679090 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:18.897765 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:19.089882 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:19.400767 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:19.589801 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:19.898545 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:20.090711 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:20.399108 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:20.589422 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:20.929298 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:21.090673 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:21.399092 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:21.589913 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:21.897999 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:22.089348 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:22.398492 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:22.590200 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:22.898262 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:23.089992 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:23.397897 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:23.588880 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:23.897230 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:24.092962 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:24.397868 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:24.589467 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:24.897916 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:25.089186 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:25.398296 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:25.590111 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:25.898190 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:26.097593 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:26.398218 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:26.589754 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:26.898561 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:27.089993 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:27.398873 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:27.589334 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:27.898081 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:28.089095 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:28.398597 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:28.590603 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:28.897834 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:29.092101 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:29.398592 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:29.590502 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:29.897929 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:30.089994 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:30.398496 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:30.589545 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:30.898576 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:31.090193 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:31.397864 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:31.589122 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:31.898341 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:32.090037 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:32.398161 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:32.590329 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:32.898608 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:33.090329 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:33.398333 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:33.589742 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:33.898627 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:34.090444 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:34.399356 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:34.589953 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0829 18:57:34.898212 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:35.089327 426808 kapi.go:107] duration metric: took 45.504479426s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0829 18:57:35.398099 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:35.897573 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:36.398587 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:36.897668 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:37.398998 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:37.897928 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:38.397299 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:38.897833 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:39.398249 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:39.898158 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:40.397946 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:40.897562 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:41.399652 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:41.897748 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:42.397861 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:42.897932 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:43.397751 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:43.897395 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:44.397680 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:44.899724 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:45.398136 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:45.898283 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:46.397638 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:46.897469 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:47.397729 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:47.897727 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:48.398340 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:48.898027 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:49.398540 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:49.898450 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:50.399497 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:50.898713 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:51.399394 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:51.899282 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:52.398332 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:52.898730 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:53.399357 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:53.899088 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:54.583660 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:54.897898 426808 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0829 18:57:55.398431 426808 kapi.go:107] duration metric: took 1m7.504622689s to wait for app.kubernetes.io/name=ingress-nginx ...
I0829 18:58:14.613355 426808 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0829 18:58:14.613380 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:15.113627 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:15.614208 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:16.113699 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:16.613894 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:17.113753 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:17.613865 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:18.114269 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:18.613283 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:19.113694 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:19.614202 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:20.113718 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:20.614014 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:21.113043 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:21.614172 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:22.113515 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:22.613771 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:23.113678 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:23.613757 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:24.113581 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:24.613517 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:25.113357 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:25.613668 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:26.113644 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:26.613468 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:27.113589 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:27.613964 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:28.114064 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:28.612977 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:29.114111 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:29.614497 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:30.113459 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:30.613524 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:31.113430 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:31.613677 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:32.113699 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:32.614248 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:33.113149 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:33.613187 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:34.113161 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:34.613073 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:35.113119 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:35.613675 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:36.113666 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:36.613748 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:37.113860 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:37.614065 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:38.113232 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:38.613262 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:39.113299 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:39.614359 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:40.113223 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:40.613659 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:41.113378 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:41.613924 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:42.115913 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:42.614132 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:43.113597 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:43.613054 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:44.112905 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:44.614063 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:45.113033 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:45.613829 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:46.114034 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:46.613328 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:47.113206 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:47.613395 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:48.113525 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:48.612967 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:49.114286 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:49.613909 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:50.113887 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:50.613864 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:51.114885 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:51.613982 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:52.114218 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:52.613592 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:53.113932 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:53.613871 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:54.113726 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:54.614467 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:55.113802 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:55.614185 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:56.113042 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:56.613437 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:57.113882 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:57.614093 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:58.113879 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:58.614359 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:59.113845 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:58:59.613617 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:00.113994 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:00.614104 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:01.114365 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:01.613521 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:02.113748 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:02.614151 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:03.113079 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:03.613178 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:04.113006 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:04.613299 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:05.113732 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:05.614155 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:06.114001 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:06.613745 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:07.113758 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:07.614005 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:08.112879 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:08.613981 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:09.114306 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:09.613872 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:10.113505 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:10.613408 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:11.114032 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:11.613363 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:12.113269 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:12.613551 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:13.113691 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:13.614000 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:14.114364 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:14.613710 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:15.113717 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:15.613788 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:16.114013 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:16.614193 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:17.113363 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:17.614005 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:18.114276 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:18.613773 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:19.114238 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:19.614199 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:20.113278 426808 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0829 18:59:20.613405 426808 kapi.go:107] duration metric: took 2m29.503248572s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0829 18:59:20.615384 426808 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-505336 cluster.
I0829 18:59:20.616849 426808 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0829 18:59:20.618227 426808 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0829 18:59:20.619712 426808 out.go:177] * Enabled addons: nvidia-device-plugin, cloud-spanner, storage-provisioner, volcano, helm-tiller, ingress-dns, metrics-server, inspektor-gadget, yakd, default-storageclass, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0829 18:59:20.620897 426808 addons.go:510] duration metric: took 2m45.286986995s for enable addons: enabled=[nvidia-device-plugin cloud-spanner storage-provisioner volcano helm-tiller ingress-dns metrics-server inspektor-gadget yakd default-storageclass volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0829 18:59:20.620941 426808 start.go:246] waiting for cluster config update ...
I0829 18:59:20.620959 426808 start.go:255] writing updated cluster config ...
I0829 18:59:20.621217 426808 ssh_runner.go:195] Run: rm -f paused
I0829 18:59:20.670441 426808 start.go:600] kubectl: 1.31.0, cluster: 1.31.0 (minor skew: 0)
I0829 18:59:20.672068 426808 out.go:177] * Done! kubectl is now configured to use "addons-505336" cluster and "default" namespace by default
==> Docker <==
Aug 29 19:08:40 addons-505336 dockerd[1340]: time="2024-08-29T19:08:40.596661177Z" level=info msg="ignoring event" container=9fa5d30490d42de5252f99c274014ef5b5e4157a936bada33bba952f1d8f788d module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:40 addons-505336 dockerd[1340]: time="2024-08-29T19:08:40.599733256Z" level=info msg="ignoring event" container=bc8552c035b7cef312339dda09dd52658efb9ad7452b0f22ad0e58410796a4ad module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:40 addons-505336 dockerd[1340]: time="2024-08-29T19:08:40.601813741Z" level=info msg="ignoring event" container=6ba3cd71e3f182ba2b42560ddbb6459e60ac17bc8daa8d18afb6b03390a7c7d9 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:40 addons-505336 dockerd[1340]: time="2024-08-29T19:08:40.678206455Z" level=info msg="ignoring event" container=cbf39dbbc3e00020b0d216ae34d16589bfd26b1c3e59cfacc4436a4e1224893d module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:40 addons-505336 dockerd[1340]: time="2024-08-29T19:08:40.843018869Z" level=info msg="ignoring event" container=e86ec71a5418a3d61e4104b6d62ff3f5ce2bec00048783a569a055534a024527 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:40 addons-505336 dockerd[1340]: time="2024-08-29T19:08:40.870928211Z" level=info msg="ignoring event" container=be1b6ca1cd14613bd2c0c01a45b5cd2b39bf547603b6d6d31a16e6e6ca0d3daa module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:40 addons-505336 dockerd[1340]: time="2024-08-29T19:08:40.909175690Z" level=info msg="ignoring event" container=c99392d19249b3fdfda3fd522076ada5590e39603bf648cc85888c77f02ca880 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:43 addons-505336 cri-dockerd[1604]: time="2024-08-29T19:08:43Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/b6cac2822e1656a911e8c3e21fb604881b212b5ba18269ef32a9698a7a769ce2/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local us-east4-a.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Aug 29 19:08:43 addons-505336 dockerd[1340]: time="2024-08-29T19:08:43.456050570Z" level=info msg="ignoring event" container=3be282e754dce4c72b5c8fc8fbf24e3f57b5592079d82725dab1ac1d67fdd1b4 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:43 addons-505336 dockerd[1340]: time="2024-08-29T19:08:43.509070226Z" level=info msg="ignoring event" container=2c68e4220999dece766e0c681bad0306335371339185c9fca8dac4564838e509 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:43 addons-505336 cri-dockerd[1604]: time="2024-08-29T19:08:43Z" level=info msg="Stop pulling image docker.io/kicbase/echo-server:1.0: Status: Downloaded newer image for kicbase/echo-server:1.0"
Aug 29 19:08:46 addons-505336 dockerd[1340]: time="2024-08-29T19:08:46.796245700Z" level=info msg="ignoring event" container=9bdb5bbbb6aa94c57a26384f85ad208053f7481fd4d529c4867a7712c33592da module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:46 addons-505336 dockerd[1340]: time="2024-08-29T19:08:46.801455636Z" level=info msg="ignoring event" container=347c39f6c0f9316409397b5abde46757f6d46601ee7506a4c13cb3fdd7fb5a1c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:46 addons-505336 dockerd[1340]: time="2024-08-29T19:08:46.970696219Z" level=info msg="ignoring event" container=1bb10db18b5cef864e5cfd4385d071b7a838ca03870e9c7b7687c3996deb49be module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:47 addons-505336 dockerd[1340]: time="2024-08-29T19:08:47.006772968Z" level=info msg="ignoring event" container=5f71bfe2caac52638fe7fa363a02b511aa1a89af359ae349a7e913ba28d18c19 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:47 addons-505336 dockerd[1340]: time="2024-08-29T19:08:47.501903626Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=d20715f8b703e129e626c12417d5431a3d29c3c8c137f8e738ee0196296d6213
Aug 29 19:08:47 addons-505336 dockerd[1340]: time="2024-08-29T19:08:47.551758741Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Aug 29 19:08:47 addons-505336 dockerd[1340]: time="2024-08-29T19:08:47.554135419Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Aug 29 19:08:47 addons-505336 dockerd[1340]: time="2024-08-29T19:08:47.562046442Z" level=info msg="ignoring event" container=d20715f8b703e129e626c12417d5431a3d29c3c8c137f8e738ee0196296d6213 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:08:47 addons-505336 dockerd[1340]: time="2024-08-29T19:08:47.703649810Z" level=info msg="ignoring event" container=d252d0d83d974e8d1134db309e9f35bdb0e60afb5e28eb9f04a162b5e99960a8 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:09:10 addons-505336 dockerd[1340]: time="2024-08-29T19:09:10.797478994Z" level=info msg="ignoring event" container=7bcb88fabc6055ae771d5d4bd718358591cd30a2c0ededbd87c7e82766baca11 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:09:11 addons-505336 dockerd[1340]: time="2024-08-29T19:09:11.281907020Z" level=info msg="ignoring event" container=865ff0e7593f3851f753ee3aa7910b933c3ca4b6bc3dca72ca2e31db85e78d4e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:09:11 addons-505336 dockerd[1340]: time="2024-08-29T19:09:11.339145398Z" level=info msg="ignoring event" container=af431d61ffaa5ef6d697a88b685cb8a38b48d1cc4c9c3ae116b8e2216ca2d52b module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:09:11 addons-505336 dockerd[1340]: time="2024-08-29T19:09:11.424428617Z" level=info msg="ignoring event" container=e90700162db654620bc235af2997bd9cd70607716de0de328f186de5141107e6 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Aug 29 19:09:11 addons-505336 dockerd[1340]: time="2024-08-29T19:09:11.488527129Z" level=info msg="ignoring event" container=834bc0d091fda50323b41f51c01d66e2322bad272880836f2d024801c8acdf4f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
c4a39b8caeee8 kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 29 seconds ago Running hello-world-app 0 b6cac2822e165 hello-world-app-55bf9c44b4-g5xvx
0ec7b927ad9f8 nginx@sha256:c04c18adc2a407740a397c8407c011fc6c90026a9b65cceddef7ae5484360158 38 seconds ago Running nginx 0 32686093fa469 nginx
a534c97fbb20a gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 1238df6a3d4b3 gcp-auth-89d5ffd79-sjw58
26b46191b2a33 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited patch 0 8c4a9e457298a ingress-nginx-admission-patch-f2kwv
f80176804ff34 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 406d636181981 ingress-nginx-admission-create-5m9dx
e1e2d5e029a8f rancher/local-path-provisioner@sha256:e34c88ae0affb1cdefbb874140d6339d4a27ec4ee420ae8199cd839997b05246 12 minutes ago Running local-path-provisioner 0 8860b3543854a local-path-provisioner-86d989889c-rhlfl
a60f0c6816401 6e38f40d628db 12 minutes ago Running storage-provisioner 0 59e9a9ed7b7b9 storage-provisioner
76912389427f4 cbb01a7bd410d 12 minutes ago Running coredns 0 865f01957f01a coredns-6f6b679f8f-2c86p
f9aa38df9c36d ad83b2ca7b09e 12 minutes ago Running kube-proxy 0 edd6c658a97bc kube-proxy-kj5d4
b64c23d78d662 604f5db92eaa8 12 minutes ago Running kube-apiserver 0 37baa47785d61 kube-apiserver-addons-505336
5d5970f46309a 045733566833c 12 minutes ago Running kube-controller-manager 0 4ec82008371f4 kube-controller-manager-addons-505336
600effe6db079 2e96e5913fc06 12 minutes ago Running etcd 0 8fc8101eb49c4 etcd-addons-505336
11536d164fd70 1766f54c897f0 12 minutes ago Running kube-scheduler 0 937c40dbc9cfc kube-scheduler-addons-505336
==> coredns [76912389427f] <==
Trace[630911734]: [30.000350004s] [30.000350004s] END
[ERROR] plugin/kubernetes: pkg/mod/k8s.io/client-go@v0.27.4/tools/cache/reflector.go:231: Failed to watch *v1.EndpointSlice: failed to list *v1.EndpointSlice: Get "https://10.96.0.1:443/apis/discovery.k8s.io/v1/endpointslices?limit=500&resourceVersion=0": dial tcp 10.96.0.1:443: i/o timeout
[INFO] plugin/kubernetes: pkg/mod/k8s.io/client-go@v0.27.4/tools/cache/reflector.go:231: failed to list *v1.Namespace: Get "https://10.96.0.1:443/api/v1/namespaces?limit=500&resourceVersion=0": dial tcp 10.96.0.1:443: i/o timeout
[INFO] plugin/kubernetes: Trace[274992665]: "Reflector ListAndWatch" name:pkg/mod/k8s.io/client-go@v0.27.4/tools/cache/reflector.go:231 (29-Aug-2024 18:56:38.388) (total time: 30000ms):
Trace[274992665]: ---"Objects listed" error:Get "https://10.96.0.1:443/api/v1/namespaces?limit=500&resourceVersion=0": dial tcp 10.96.0.1:443: i/o timeout 30000ms (18:57:08.389)
Trace[274992665]: [30.000347993s] [30.000347993s] END
[ERROR] plugin/kubernetes: pkg/mod/k8s.io/client-go@v0.27.4/tools/cache/reflector.go:231: Failed to watch *v1.Namespace: failed to list *v1.Namespace: Get "https://10.96.0.1:443/api/v1/namespaces?limit=500&resourceVersion=0": dial tcp 10.96.0.1:443: i/o timeout
[INFO] Reloading
[INFO] plugin/reload: Running configuration SHA512 = 05e3eaddc414b2d71a69b2e2bc6f2681fc1f4d04bcdd3acc1a41457bb7db518208b95ddfc4c9fffedc59c25a8faf458be1af4915a4a3c0d6777cb7a346bc5d86
[INFO] Reloading complete
[INFO] 127.0.0.1:59405 - 21872 "HINFO IN 764184260452644925.6571257308608008398. udp 56 false 512" NXDOMAIN qr,rd,ra 56 0.009208319s
[INFO] 10.244.0.26:52465 - 61767 "A IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000405931s
[INFO] 10.244.0.26:36476 - 65111 "AAAA IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000502619s
[INFO] 10.244.0.26:55198 - 38137 "A IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000123193s
[INFO] 10.244.0.26:38333 - 9663 "AAAA IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000152873s
[INFO] 10.244.0.26:36202 - 39299 "AAAA IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000119616s
[INFO] 10.244.0.26:37698 - 13537 "A IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000188924s
[INFO] 10.244.0.26:56550 - 39640 "A IN storage.googleapis.com.us-east4-a.c.k8s-minikube.internal. udp 86 false 1232" NXDOMAIN qr,rd,ra 75 0.00726122s
[INFO] 10.244.0.26:60563 - 9991 "AAAA IN storage.googleapis.com.us-east4-a.c.k8s-minikube.internal. udp 86 false 1232" NXDOMAIN qr,rd,ra 75 0.008000464s
[INFO] 10.244.0.26:37248 - 30073 "A IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.00582152s
[INFO] 10.244.0.26:33710 - 38070 "AAAA IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.005997874s
[INFO] 10.244.0.26:53380 - 401 "A IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.004069723s
[INFO] 10.244.0.26:40918 - 60520 "AAAA IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.004557015s
[INFO] 10.244.0.26:50915 - 14253 "AAAA IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 240 0.000950356s
[INFO] 10.244.0.26:56831 - 52255 "A IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 458 0.001094926s
==> describe nodes <==
Name: addons-505336
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-505336
kubernetes.io/os=linux
minikube.k8s.io/commit=5512bd76519cf55fa04aeca1cd01a1369e298033
minikube.k8s.io/name=addons-505336
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_08_29T18_56_30_0700
minikube.k8s.io/version=v1.33.1
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-505336
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Thu, 29 Aug 2024 18:56:27 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-505336
AcquireTime: <unset>
RenewTime: Thu, 29 Aug 2024 19:09:05 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Thu, 29 Aug 2024 19:09:05 +0000 Thu, 29 Aug 2024 18:56:25 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Thu, 29 Aug 2024 19:09:05 +0000 Thu, 29 Aug 2024 18:56:25 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Thu, 29 Aug 2024 19:09:05 +0000 Thu, 29 Aug 2024 18:56:25 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Thu, 29 Aug 2024 19:09:05 +0000 Thu, 29 Aug 2024 18:56:27 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-505336
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
System Info:
Machine ID: 25840fc7878c4172be50473ab8518e0b
System UUID: ad351877-c4e8-4d9a-8fdb-bd3c50cb3f4a
Boot ID: 8d049dc3-d201-4992-9948-d4c3816a3020
Kernel Version: 5.15.0-1067-gcp
OS Image: Ubuntu 22.04.4 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.2.0
Kubelet Version: v1.31.0
Kube-Proxy Version:
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (12 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m14s
default hello-world-app-55bf9c44b4-g5xvx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 30s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 39s
gcp-auth gcp-auth-89d5ffd79-sjw58 0 (0%) 0 (0%) 0 (0%) 0 (0%) 10m
kube-system coredns-6f6b679f8f-2c86p 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-505336 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-505336 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-505336 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-kj5d4 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-505336 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
local-path-storage local-path-provisioner-86d989889c-rhlfl 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-505336 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-505336 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-505336 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-505336 event: Registered Node addons-505336 in Controller
==> dmesg <==
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 7a 60 3b 9f ac e7 08 06
[ +1.384095] IPv4: martian source 10.244.0.1 from 10.244.0.17, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 5a d8 23 28 c3 f4 08 06
[ +1.301152] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 3e ed 65 be af e3 08 06
[ +5.893875] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 3e 43 8f a8 7f 7e 08 06
[ +0.096252] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff a2 c5 c2 03 e0 f5 08 06
[ +0.076818] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff 56 49 69 8b fb c1 08 06
[ +22.440993] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff da 90 14 68 71 81 08 06
[Aug29 18:58] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 2e a5 d9 41 7f 44 08 06
[ +0.032870] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff ca 75 87 14 46 84 08 06
[Aug29 18:59] IPv4: martian source 10.244.0.1 from 10.244.0.26, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 5e 9c a6 72 9c fc 08 06
[ +0.000574] IPv4: martian source 10.244.0.26 from 10.244.0.2, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff e6 17 79 01 18 43 08 06
[Aug29 19:08] IPv4: martian source 10.244.0.1 from 10.244.0.29, on dev eth0
[ +0.000010] ll header: 00000000: ff ff ff ff ff ff 3e ce 45 e2 8e 3d 08 06
[ +34.846395] IPv4: martian source 10.244.0.37 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff da 90 14 68 71 81 08 06
==> etcd [600effe6db07] <==
{"level":"info","ts":"2024-08-29T18:56:25.787598Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-08-29T18:56:25.787620Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-08-29T18:56:25.787900Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-08-29T18:56:25.787904Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-08-29T18:56:25.787925Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-08-29T18:56:25.787973Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-08-29T18:56:25.787995Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-08-29T18:56:25.788724Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-08-29T18:56:25.788761Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-08-29T18:56:25.789510Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-08-29T18:56:25.789932Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-08-29T18:56:46.991268Z","caller":"traceutil/trace.go:171","msg":"trace[2070930315] transaction","detail":"{read_only:false; response_revision:808; number_of_response:1; }","duration":"102.145389ms","start":"2024-08-29T18:56:46.889101Z","end":"2024-08-29T18:56:46.991246Z","steps":["trace[2070930315] 'process raft request' (duration: 100.591885ms)"],"step_count":1}
{"level":"info","ts":"2024-08-29T18:56:46.992663Z","caller":"traceutil/trace.go:171","msg":"trace[746986554] transaction","detail":"{read_only:false; response_revision:809; number_of_response:1; }","duration":"101.504246ms","start":"2024-08-29T18:56:46.891140Z","end":"2024-08-29T18:56:46.992644Z","steps":["trace[746986554] 'process raft request' (duration: 101.004918ms)"],"step_count":1}
{"level":"info","ts":"2024-08-29T18:56:46.992801Z","caller":"traceutil/trace.go:171","msg":"trace[206749501] transaction","detail":"{read_only:false; response_revision:810; number_of_response:1; }","duration":"101.532069ms","start":"2024-08-29T18:56:46.891259Z","end":"2024-08-29T18:56:46.992791Z","steps":["trace[206749501] 'process raft request' (duration: 100.973944ms)"],"step_count":1}
{"level":"info","ts":"2024-08-29T18:56:46.992947Z","caller":"traceutil/trace.go:171","msg":"trace[1518821797] transaction","detail":"{read_only:false; response_revision:811; number_of_response:1; }","duration":"101.517405ms","start":"2024-08-29T18:56:46.891422Z","end":"2024-08-29T18:56:46.992939Z","steps":["trace[1518821797] 'process raft request' (duration: 100.86305ms)"],"step_count":1}
{"level":"warn","ts":"2024-08-29T18:57:15.272337Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"159.93534ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-08-29T18:57:15.272477Z","caller":"traceutil/trace.go:171","msg":"trace[295395045] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:1070; }","duration":"160.101553ms","start":"2024-08-29T18:57:15.112358Z","end":"2024-08-29T18:57:15.272460Z","steps":["trace[295395045] 'range keys from in-memory index tree' (duration: 159.867036ms)"],"step_count":1}
{"level":"warn","ts":"2024-08-29T18:57:54.580717Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"179.087676ms","expected-duration":"100ms","prefix":"","request":"header:<ID:8128031541713176987 > lease_revoke:<id:70cc919f7efc08dd>","response":"size:29"}
{"level":"info","ts":"2024-08-29T18:57:54.580899Z","caller":"traceutil/trace.go:171","msg":"trace[1785589315] transaction","detail":"{read_only:false; response_revision:1270; number_of_response:1; }","duration":"174.509997ms","start":"2024-08-29T18:57:54.406379Z","end":"2024-08-29T18:57:54.580889Z","steps":["trace[1785589315] 'process raft request' (duration: 174.416555ms)"],"step_count":1}
{"level":"info","ts":"2024-08-29T18:57:54.580918Z","caller":"traceutil/trace.go:171","msg":"trace[1988155826] linearizableReadLoop","detail":"{readStateIndex:1303; appliedIndex:1302; }","duration":"185.110299ms","start":"2024-08-29T18:57:54.395783Z","end":"2024-08-29T18:57:54.580893Z","steps":["trace[1988155826] 'read index received' (duration: 5.695664ms)","trace[1988155826] 'applied index is now lower than readState.Index' (duration: 179.411618ms)"],"step_count":2}
{"level":"warn","ts":"2024-08-29T18:57:54.581149Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"185.343194ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-08-29T18:57:54.581196Z","caller":"traceutil/trace.go:171","msg":"trace[1796819428] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:1270; }","duration":"185.406877ms","start":"2024-08-29T18:57:54.395778Z","end":"2024-08-29T18:57:54.581185Z","steps":["trace[1796819428] 'agreement among raft nodes before linearized reading' (duration: 185.254384ms)"],"step_count":1}
{"level":"info","ts":"2024-08-29T19:06:25.825015Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1885}
{"level":"info","ts":"2024-08-29T19:06:25.859286Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1885,"took":"33.722495ms","hash":3218043340,"current-db-size-bytes":8724480,"current-db-size":"8.7 MB","current-db-size-in-use-bytes":4857856,"current-db-size-in-use":"4.9 MB"}
{"level":"info","ts":"2024-08-29T19:06:25.859334Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":3218043340,"revision":1885,"compact-revision":-1}
==> gcp-auth [a534c97fbb20] <==
2024/08/29 18:59:58 Ready to write response ...
2024/08/29 19:08:05 Ready to marshal response ...
2024/08/29 19:08:05 Ready to write response ...
2024/08/29 19:08:06 Ready to marshal response ...
2024/08/29 19:08:06 Ready to write response ...
2024/08/29 19:08:06 Ready to marshal response ...
2024/08/29 19:08:06 Ready to write response ...
2024/08/29 19:08:07 Ready to marshal response ...
2024/08/29 19:08:07 Ready to write response ...
2024/08/29 19:08:10 Ready to marshal response ...
2024/08/29 19:08:10 Ready to write response ...
2024/08/29 19:08:14 Ready to marshal response ...
2024/08/29 19:08:14 Ready to write response ...
2024/08/29 19:08:26 Ready to marshal response ...
2024/08/29 19:08:26 Ready to write response ...
2024/08/29 19:08:26 Ready to marshal response ...
2024/08/29 19:08:26 Ready to write response ...
2024/08/29 19:08:26 Ready to marshal response ...
2024/08/29 19:08:26 Ready to write response ...
2024/08/29 19:08:31 Ready to marshal response ...
2024/08/29 19:08:31 Ready to write response ...
2024/08/29 19:08:33 Ready to marshal response ...
2024/08/29 19:08:33 Ready to write response ...
2024/08/29 19:08:42 Ready to marshal response ...
2024/08/29 19:08:42 Ready to write response ...
==> kernel <==
19:09:12 up 21:51, 0 users, load average: 0.88, 0.51, 0.35
Linux addons-505336 5.15.0-1067-gcp #75~20.04.1-Ubuntu SMP Wed Aug 7 20:43:22 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.4 LTS"
==> kube-apiserver [b64c23d78d66] <==
W0829 18:59:50.085001 1 cacher.go:171] Terminating all watchers from cacher numatopologies.nodeinfo.volcano.sh
W0829 18:59:50.481430 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0829 18:59:50.595730 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0829 18:59:50.994697 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0829 19:08:14.931399 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0829 19:08:15.916090 1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
I0829 19:08:26.575615 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.99.91.187"}
I0829 19:08:27.581443 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0829 19:08:28.600616 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0829 19:08:32.960035 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0829 19:08:33.123425 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.105.202.92"}
I0829 19:08:42.692729 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.108.2.228"}
I0829 19:08:46.662028 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0829 19:08:46.662082 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0829 19:08:46.674317 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0829 19:08:46.674423 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0829 19:08:46.675352 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0829 19:08:46.675385 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0829 19:08:46.685014 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0829 19:08:46.685063 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0829 19:08:46.695427 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0829 19:08:46.695475 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0829 19:08:47.676287 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W0829 19:08:47.696566 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
W0829 19:08:47.704547 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
==> kube-controller-manager [5d5970f46309] <==
E0829 19:08:55.960195 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0829 19:08:56.022136 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0829 19:08:56.022174 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0829 19:08:59.343084 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0829 19:08:59.343123 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0829 19:09:00.060971 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0829 19:09:00.061019 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0829 19:09:00.967705 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0829 19:09:00.967747 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0829 19:09:01.342704 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0829 19:09:01.342743 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0829 19:09:02.419575 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0829 19:09:02.419616 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0829 19:09:04.799559 1 shared_informer.go:313] Waiting for caches to sync for resource quota
I0829 19:09:04.799592 1 shared_informer.go:320] Caches are synced for resource quota
I0829 19:09:04.910604 1 shared_informer.go:313] Waiting for caches to sync for garbage collector
I0829 19:09:04.910650 1 shared_informer.go:320] Caches are synced for garbage collector
I0829 19:09:05.286713 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-505336"
W0829 19:09:06.254234 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0829 19:09:06.254273 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0829 19:09:08.004724 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0829 19:09:08.004771 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0829 19:09:10.383697 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0829 19:09:10.383736 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0829 19:09:11.221111 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-6fb4cdfc84" duration="9.35µs"
==> kube-proxy [f9aa38df9c36] <==
I0829 18:56:37.291945 1 server_linux.go:66] "Using iptables proxy"
I0829 18:56:38.094182 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0829 18:56:38.094269 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0829 18:56:38.593375 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0829 18:56:38.593442 1 server_linux.go:169] "Using iptables Proxier"
I0829 18:56:38.595874 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0829 18:56:38.596298 1 server.go:483] "Version info" version="v1.31.0"
I0829 18:56:38.596314 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0829 18:56:38.598409 1 config.go:197] "Starting service config controller"
I0829 18:56:38.598427 1 shared_informer.go:313] Waiting for caches to sync for service config
I0829 18:56:38.598449 1 config.go:104] "Starting endpoint slice config controller"
I0829 18:56:38.598455 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0829 18:56:38.675671 1 config.go:326] "Starting node config controller"
I0829 18:56:38.675690 1 shared_informer.go:313] Waiting for caches to sync for node config
I0829 18:56:38.775045 1 shared_informer.go:320] Caches are synced for service config
I0829 18:56:38.775043 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0829 18:56:38.776590 1 shared_informer.go:320] Caches are synced for node config
==> kube-scheduler [11536d164fd7] <==
W0829 18:56:27.399473 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0829 18:56:27.399497 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.216435 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0829 18:56:28.216479 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.220757 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0829 18:56:28.220792 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.252070 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0829 18:56:28.252112 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.259348 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0829 18:56:28.259377 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.300690 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0829 18:56:28.300722 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.360101 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0829 18:56:28.360148 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.374495 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0829 18:56:28.374546 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.448389 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0829 18:56:28.448441 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.530011 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0829 18:56:28.530057 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.548389 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0829 18:56:28.548432 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0829 18:56:28.549094 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0829 18:56:28.549123 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
I0829 18:56:28.797115 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Aug 29 19:08:49 addons-505336 kubelet[2438]: I0829 19:08:49.498098 2438 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fbc68ee-15f3-4427-a51a-61672bef2410" path="/var/lib/kubelet/pods/9fbc68ee-15f3-4427-a51a-61672bef2410/volumes"
Aug 29 19:08:51 addons-505336 kubelet[2438]: E0829 19:08:51.489732 2438 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="f9b774b1-c20c-4cf3-814a-6d65d22b20d7"
Aug 29 19:09:01 addons-505336 kubelet[2438]: E0829 19:09:01.489403 2438 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox\\\"\"" pod="default/registry-test" podUID="ce9ebc43-07ee-4c49-bbf9-3a21a2b71153"
Aug 29 19:09:05 addons-505336 kubelet[2438]: E0829 19:09:05.489648 2438 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="f9b774b1-c20c-4cf3-814a-6d65d22b20d7"
Aug 29 19:09:10 addons-505336 kubelet[2438]: I0829 19:09:10.966110 2438 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/ce9ebc43-07ee-4c49-bbf9-3a21a2b71153-gcp-creds\") pod \"ce9ebc43-07ee-4c49-bbf9-3a21a2b71153\" (UID: \"ce9ebc43-07ee-4c49-bbf9-3a21a2b71153\") "
Aug 29 19:09:10 addons-505336 kubelet[2438]: I0829 19:09:10.966173 2438 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk8rs\" (UniqueName: \"kubernetes.io/projected/ce9ebc43-07ee-4c49-bbf9-3a21a2b71153-kube-api-access-vk8rs\") pod \"ce9ebc43-07ee-4c49-bbf9-3a21a2b71153\" (UID: \"ce9ebc43-07ee-4c49-bbf9-3a21a2b71153\") "
Aug 29 19:09:10 addons-505336 kubelet[2438]: I0829 19:09:10.966236 2438 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce9ebc43-07ee-4c49-bbf9-3a21a2b71153-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "ce9ebc43-07ee-4c49-bbf9-3a21a2b71153" (UID: "ce9ebc43-07ee-4c49-bbf9-3a21a2b71153"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Aug 29 19:09:10 addons-505336 kubelet[2438]: I0829 19:09:10.968302 2438 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce9ebc43-07ee-4c49-bbf9-3a21a2b71153-kube-api-access-vk8rs" (OuterVolumeSpecName: "kube-api-access-vk8rs") pod "ce9ebc43-07ee-4c49-bbf9-3a21a2b71153" (UID: "ce9ebc43-07ee-4c49-bbf9-3a21a2b71153"). InnerVolumeSpecName "kube-api-access-vk8rs". PluginName "kubernetes.io/projected", VolumeGidValue ""
Aug 29 19:09:11 addons-505336 kubelet[2438]: I0829 19:09:11.066744 2438 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-vk8rs\" (UniqueName: \"kubernetes.io/projected/ce9ebc43-07ee-4c49-bbf9-3a21a2b71153-kube-api-access-vk8rs\") on node \"addons-505336\" DevicePath \"\""
Aug 29 19:09:11 addons-505336 kubelet[2438]: I0829 19:09:11.066804 2438 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/ce9ebc43-07ee-4c49-bbf9-3a21a2b71153-gcp-creds\") on node \"addons-505336\" DevicePath \"\""
Aug 29 19:09:11 addons-505336 kubelet[2438]: I0829 19:09:11.495230 2438 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce9ebc43-07ee-4c49-bbf9-3a21a2b71153" path="/var/lib/kubelet/pods/ce9ebc43-07ee-4c49-bbf9-3a21a2b71153/volumes"
Aug 29 19:09:11 addons-505336 kubelet[2438]: I0829 19:09:11.578054 2438 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs5bp\" (UniqueName: \"kubernetes.io/projected/0bc8f454-eced-450f-ab5d-b961648307b9-kube-api-access-gs5bp\") pod \"0bc8f454-eced-450f-ab5d-b961648307b9\" (UID: \"0bc8f454-eced-450f-ab5d-b961648307b9\") "
Aug 29 19:09:11 addons-505336 kubelet[2438]: I0829 19:09:11.578098 2438 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp8fx\" (UniqueName: \"kubernetes.io/projected/ef74c371-cbb9-4c81-91dd-dcbc748f81d0-kube-api-access-xp8fx\") pod \"ef74c371-cbb9-4c81-91dd-dcbc748f81d0\" (UID: \"ef74c371-cbb9-4c81-91dd-dcbc748f81d0\") "
Aug 29 19:09:11 addons-505336 kubelet[2438]: I0829 19:09:11.579868 2438 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef74c371-cbb9-4c81-91dd-dcbc748f81d0-kube-api-access-xp8fx" (OuterVolumeSpecName: "kube-api-access-xp8fx") pod "ef74c371-cbb9-4c81-91dd-dcbc748f81d0" (UID: "ef74c371-cbb9-4c81-91dd-dcbc748f81d0"). InnerVolumeSpecName "kube-api-access-xp8fx". PluginName "kubernetes.io/projected", VolumeGidValue ""
Aug 29 19:09:11 addons-505336 kubelet[2438]: I0829 19:09:11.579870 2438 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc8f454-eced-450f-ab5d-b961648307b9-kube-api-access-gs5bp" (OuterVolumeSpecName: "kube-api-access-gs5bp") pod "0bc8f454-eced-450f-ab5d-b961648307b9" (UID: "0bc8f454-eced-450f-ab5d-b961648307b9"). InnerVolumeSpecName "kube-api-access-gs5bp". PluginName "kubernetes.io/projected", VolumeGidValue ""
Aug 29 19:09:11 addons-505336 kubelet[2438]: I0829 19:09:11.679475 2438 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-gs5bp\" (UniqueName: \"kubernetes.io/projected/0bc8f454-eced-450f-ab5d-b961648307b9-kube-api-access-gs5bp\") on node \"addons-505336\" DevicePath \"\""
Aug 29 19:09:11 addons-505336 kubelet[2438]: I0829 19:09:11.679516 2438 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-xp8fx\" (UniqueName: \"kubernetes.io/projected/ef74c371-cbb9-4c81-91dd-dcbc748f81d0-kube-api-access-xp8fx\") on node \"addons-505336\" DevicePath \"\""
Aug 29 19:09:12 addons-505336 kubelet[2438]: I0829 19:09:12.052768 2438 scope.go:117] "RemoveContainer" containerID="af431d61ffaa5ef6d697a88b685cb8a38b48d1cc4c9c3ae116b8e2216ca2d52b"
Aug 29 19:09:12 addons-505336 kubelet[2438]: I0829 19:09:12.069149 2438 scope.go:117] "RemoveContainer" containerID="af431d61ffaa5ef6d697a88b685cb8a38b48d1cc4c9c3ae116b8e2216ca2d52b"
Aug 29 19:09:12 addons-505336 kubelet[2438]: E0829 19:09:12.071321 2438 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: af431d61ffaa5ef6d697a88b685cb8a38b48d1cc4c9c3ae116b8e2216ca2d52b" containerID="af431d61ffaa5ef6d697a88b685cb8a38b48d1cc4c9c3ae116b8e2216ca2d52b"
Aug 29 19:09:12 addons-505336 kubelet[2438]: I0829 19:09:12.071360 2438 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"af431d61ffaa5ef6d697a88b685cb8a38b48d1cc4c9c3ae116b8e2216ca2d52b"} err="failed to get container status \"af431d61ffaa5ef6d697a88b685cb8a38b48d1cc4c9c3ae116b8e2216ca2d52b\": rpc error: code = Unknown desc = Error response from daemon: No such container: af431d61ffaa5ef6d697a88b685cb8a38b48d1cc4c9c3ae116b8e2216ca2d52b"
Aug 29 19:09:12 addons-505336 kubelet[2438]: I0829 19:09:12.071378 2438 scope.go:117] "RemoveContainer" containerID="865ff0e7593f3851f753ee3aa7910b933c3ca4b6bc3dca72ca2e31db85e78d4e"
Aug 29 19:09:12 addons-505336 kubelet[2438]: I0829 19:09:12.086555 2438 scope.go:117] "RemoveContainer" containerID="865ff0e7593f3851f753ee3aa7910b933c3ca4b6bc3dca72ca2e31db85e78d4e"
Aug 29 19:09:12 addons-505336 kubelet[2438]: E0829 19:09:12.087325 2438 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 865ff0e7593f3851f753ee3aa7910b933c3ca4b6bc3dca72ca2e31db85e78d4e" containerID="865ff0e7593f3851f753ee3aa7910b933c3ca4b6bc3dca72ca2e31db85e78d4e"
Aug 29 19:09:12 addons-505336 kubelet[2438]: I0829 19:09:12.087365 2438 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"865ff0e7593f3851f753ee3aa7910b933c3ca4b6bc3dca72ca2e31db85e78d4e"} err="failed to get container status \"865ff0e7593f3851f753ee3aa7910b933c3ca4b6bc3dca72ca2e31db85e78d4e\": rpc error: code = Unknown desc = Error response from daemon: No such container: 865ff0e7593f3851f753ee3aa7910b933c3ca4b6bc3dca72ca2e31db85e78d4e"
==> storage-provisioner [a60f0c681640] <==
I0829 18:56:42.281135 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0829 18:56:42.386247 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0829 18:56:42.386296 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0829 18:56:42.484026 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0829 18:56:42.484376 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-505336_60e9b849-e378-4aed-b33d-e0e90a4d1d41!
I0829 18:56:42.486002 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"ac97d5f2-4185-4dd9-b5ca-de8119bbf3dc", APIVersion:"v1", ResourceVersion:"552", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-505336_60e9b849-e378-4aed-b33d-e0e90a4d1d41 became leader
I0829 18:56:42.585378 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-505336_60e9b849-e378-4aed-b33d-e0e90a4d1d41!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-505336 -n addons-505336
helpers_test.go:261: (dbg) Run: kubectl --context addons-505336 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-505336 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-505336 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-505336/192.168.49.2
Start Time: Thu, 29 Aug 2024 18:59:58 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.28
IPs:
IP: 10.244.0.28
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-zw2tc (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-zw2tc:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m14s default-scheduler Successfully assigned default/busybox to addons-505336
Normal Pulling 7m43s (x4 over 9m14s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m43s (x4 over 9m14s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m43s (x4 over 9m14s) kubelet Error: ErrImagePull
Warning Failed 7m31s (x6 over 9m13s) kubelet Error: ImagePullBackOff
Normal BackOff 4m9s (x20 over 9m13s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (72.39s)