=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:328: registry stabilized in 1.985626ms
addons_test.go:330: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-vmjl4" [83533b4a-57e0-46f9-8a17-5cbf2c880aea] Running
addons_test.go:330: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.002349952s
addons_test.go:333: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-n8nl5" [60333ce0-1ed5-49fd-bf03-d4ba227481ce] Running
addons_test.go:333: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.003465091s
addons_test.go:338: (dbg) Run: kubectl --context addons-928332 delete po -l run=registry-test --now
addons_test.go:343: (dbg) Run: kubectl --context addons-928332 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:343: (dbg) Non-zero exit: kubectl --context addons-928332 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.076969725s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:345: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-928332 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:349: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:357: (dbg) Run: out/minikube-linux-amd64 -p addons-928332 ip
2024/09/20 17:40:24 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:386: (dbg) Run: out/minikube-linux-amd64 -p addons-928332 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-928332
helpers_test.go:235: (dbg) docker inspect addons-928332:
-- stdout --
[
{
"Id": "9b5691e6d8518264b46346aecb82b15757936a4517afb7fa259da455f99a8d51",
"Created": "2024-09-20T17:27:29.276796394Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 536773,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-20T17:27:29.373033108Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:bb3bcbaabeeeadbf6b43ae7d1d07e504b3c8a94ec024df89bcb237eba4f5e9b3",
"ResolvConfPath": "/var/lib/docker/containers/9b5691e6d8518264b46346aecb82b15757936a4517afb7fa259da455f99a8d51/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/9b5691e6d8518264b46346aecb82b15757936a4517afb7fa259da455f99a8d51/hostname",
"HostsPath": "/var/lib/docker/containers/9b5691e6d8518264b46346aecb82b15757936a4517afb7fa259da455f99a8d51/hosts",
"LogPath": "/var/lib/docker/containers/9b5691e6d8518264b46346aecb82b15757936a4517afb7fa259da455f99a8d51/9b5691e6d8518264b46346aecb82b15757936a4517afb7fa259da455f99a8d51-json.log",
"Name": "/addons-928332",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-928332:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-928332",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/f31b9d12cc25b6a780fdeb91eeb6a97c9a410eb0023d16b25453909130c6a75c-init/diff:/var/lib/docker/overlay2/2c41c746a7e9a13c324bcaad7728a158b54385725f3a4f27dce1d7d47e947b02/diff",
"MergedDir": "/var/lib/docker/overlay2/f31b9d12cc25b6a780fdeb91eeb6a97c9a410eb0023d16b25453909130c6a75c/merged",
"UpperDir": "/var/lib/docker/overlay2/f31b9d12cc25b6a780fdeb91eeb6a97c9a410eb0023d16b25453909130c6a75c/diff",
"WorkDir": "/var/lib/docker/overlay2/f31b9d12cc25b6a780fdeb91eeb6a97c9a410eb0023d16b25453909130c6a75c/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-928332",
"Source": "/var/lib/docker/volumes/addons-928332/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-928332",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-928332",
"name.minikube.sigs.k8s.io": "addons-928332",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "d5ab859e1995dab7f18ae242d02a746c5ed54e654c2535fce959d5595d23061f",
"SandboxKey": "/var/run/docker/netns/d5ab859e1995",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33164"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33165"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33168"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33166"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "33167"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-928332": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "19023f0cc3b9748cc2058cd655d2e57438876efaa83ebc52db9599dac9a2a6b1",
"EndpointID": "8ea6883d026a190c4300466882c0a80d1c00f8a61342caafa28b9b4534ef1025",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-928332",
"9b5691e6d851"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-928332 -n addons-928332
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-928332 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-amd64 -p addons-928332 logs -n 25: (1.054507098s)
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-only-257357 | download-only-257357 | jenkins | v1.34.0 | 20 Sep 24 17:27 UTC | 20 Sep 24 17:27 UTC |
| start | --download-only -p | download-docker-014158 | jenkins | v1.34.0 | 20 Sep 24 17:27 UTC | |
| | download-docker-014158 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p download-docker-014158 | download-docker-014158 | jenkins | v1.34.0 | 20 Sep 24 17:27 UTC | 20 Sep 24 17:27 UTC |
| start | --download-only -p | binary-mirror-234390 | jenkins | v1.34.0 | 20 Sep 24 17:27 UTC | |
| | binary-mirror-234390 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:45189 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-234390 | binary-mirror-234390 | jenkins | v1.34.0 | 20 Sep 24 17:27 UTC | 20 Sep 24 17:27 UTC |
| addons | enable dashboard -p | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:27 UTC | |
| | addons-928332 | | | | | |
| addons | disable dashboard -p | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:27 UTC | |
| | addons-928332 | | | | | |
| start | -p addons-928332 --wait=true | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:27 UTC | 20 Sep 24 17:30 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| addons | addons-928332 addons disable | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:31 UTC | 20 Sep 24 17:31 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | enable headlamp | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| | -p addons-928332 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-928332 addons disable | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| | -p addons-928332 | | | | | |
| addons | addons-928332 addons disable | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ssh | addons-928332 ssh cat | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| | /opt/local-path-provisioner/pvc-2ae5ff3b-f3ce-4447-8457-6ba785ff59ec_default_test-pvc/file1 | | | | | |
| addons | addons-928332 addons disable | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:40 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable cloud-spanner -p | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| | addons-928332 | | | | | |
| addons | addons-928332 addons | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| | addons-928332 | | | | | |
| ssh | addons-928332 ssh curl -s | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-928332 ip | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| addons | addons-928332 addons disable | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:39 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-928332 addons disable | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:39 UTC | 20 Sep 24 17:40 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| addons | addons-928332 addons | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:40 UTC | |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-928332 ip | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:40 UTC | 20 Sep 24 17:40 UTC |
| addons | addons-928332 addons disable | addons-928332 | jenkins | v1.34.0 | 20 Sep 24 17:40 UTC | 20 Sep 24 17:40 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/20 17:27:06
Running on machine: ubuntu-20-agent-14
Binary: Built with gc go1.23.0 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0920 17:27:06.446722 536038 out.go:345] Setting OutFile to fd 1 ...
I0920 17:27:06.446821 536038 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0920 17:27:06.446829 536038 out.go:358] Setting ErrFile to fd 2...
I0920 17:27:06.446834 536038 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0920 17:27:06.447022 536038 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19679-527899/.minikube/bin
I0920 17:27:06.447523 536038 out.go:352] Setting JSON to false
I0920 17:27:06.448503 536038 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":4178,"bootTime":1726849048,"procs":305,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1069-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0920 17:27:06.448605 536038 start.go:139] virtualization: kvm guest
I0920 17:27:06.450587 536038 out.go:177] * [addons-928332] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0920 17:27:06.451798 536038 out.go:177] - MINIKUBE_LOCATION=19679
I0920 17:27:06.451833 536038 notify.go:220] Checking for updates...
I0920 17:27:06.453971 536038 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0920 17:27:06.455023 536038 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19679-527899/kubeconfig
I0920 17:27:06.456066 536038 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19679-527899/.minikube
I0920 17:27:06.457062 536038 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0920 17:27:06.458047 536038 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0920 17:27:06.459159 536038 driver.go:394] Setting default libvirt URI to qemu:///system
I0920 17:27:06.481115 536038 docker.go:123] docker version: linux-27.3.0:Docker Engine - Community
I0920 17:27:06.481249 536038 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0920 17:27:06.528122 536038 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:30 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-20 17:27:06.519352896 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.3.0 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:[WARNING: br
idge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled] ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.17.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.6] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0920 17:27:06.528283 536038 docker.go:318] overlay module found
I0920 17:27:06.530368 536038 out.go:177] * Using the docker driver based on user configuration
I0920 17:27:06.531260 536038 start.go:297] selected driver: docker
I0920 17:27:06.531270 536038 start.go:901] validating driver "docker" against <nil>
I0920 17:27:06.531281 536038 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0920 17:27:06.532033 536038 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0920 17:27:06.574355 536038 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:false BridgeNfIP6Tables:false Debug:false NFd:30 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-20 17:27:06.565177934 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x
86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.3.0 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:[WARNING: br
idge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled] ServerErrors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.17.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.6] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0920 17:27:06.574546 536038 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0920 17:27:06.574795 536038 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0920 17:27:06.576215 536038 out.go:177] * Using Docker driver with root privileges
I0920 17:27:06.577177 536038 cni.go:84] Creating CNI manager for ""
I0920 17:27:06.577267 536038 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0920 17:27:06.577282 536038 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0920 17:27:06.577363 536038 start.go:340] cluster config:
{Name:addons-928332 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-928332 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0920 17:27:06.578482 536038 out.go:177] * Starting "addons-928332" primary control-plane node in "addons-928332" cluster
I0920 17:27:06.579351 536038 cache.go:121] Beginning downloading kic base image for docker with docker
I0920 17:27:06.580306 536038 out.go:177] * Pulling base image v0.0.45-1726589491-19662 ...
I0920 17:27:06.581188 536038 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0920 17:27:06.581217 536038 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 in local docker daemon
I0920 17:27:06.581223 536038 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19679-527899/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4
I0920 17:27:06.581264 536038 cache.go:56] Caching tarball of preloaded images
I0920 17:27:06.581376 536038 preload.go:172] Found /home/jenkins/minikube-integration/19679-527899/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0920 17:27:06.581433 536038 cache.go:59] Finished verifying existence of preloaded tar for v1.31.1 on docker
I0920 17:27:06.581747 536038 profile.go:143] Saving config to /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/config.json ...
I0920 17:27:06.581770 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/config.json: {Name:mk7f3871b67d02be1e581dd25498cbb4af4547e0 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:06.595307 536038 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 to local cache
I0920 17:27:06.595401 536038 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 in local cache directory
I0920 17:27:06.595414 536038 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 in local cache directory, skipping pull
I0920 17:27:06.595419 536038 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 exists in cache, skipping pull
I0920 17:27:06.595426 536038 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 as a tarball
I0920 17:27:06.595432 536038 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 from local cache
I0920 17:27:18.222675 536038 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 from cached tarball
I0920 17:27:18.222735 536038 cache.go:194] Successfully downloaded all kic artifacts
I0920 17:27:18.222789 536038 start.go:360] acquireMachinesLock for addons-928332: {Name:mk81c136aed197d9a3619f5b5b901c1b880f960a Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0920 17:27:18.222904 536038 start.go:364] duration metric: took 90.112µs to acquireMachinesLock for "addons-928332"
I0920 17:27:18.222938 536038 start.go:93] Provisioning new machine with config: &{Name:addons-928332 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-928332 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0920 17:27:18.223045 536038 start.go:125] createHost starting for "" (driver="docker")
I0920 17:27:18.228449 536038 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0920 17:27:18.228706 536038 start.go:159] libmachine.API.Create for "addons-928332" (driver="docker")
I0920 17:27:18.228757 536038 client.go:168] LocalClient.Create starting
I0920 17:27:18.228854 536038 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19679-527899/.minikube/certs/ca.pem
I0920 17:27:18.331090 536038 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19679-527899/.minikube/certs/cert.pem
I0920 17:27:18.415483 536038 cli_runner.go:164] Run: docker network inspect addons-928332 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0920 17:27:18.430561 536038 cli_runner.go:211] docker network inspect addons-928332 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0920 17:27:18.430632 536038 network_create.go:284] running [docker network inspect addons-928332] to gather additional debugging logs...
I0920 17:27:18.430654 536038 cli_runner.go:164] Run: docker network inspect addons-928332
W0920 17:27:18.444961 536038 cli_runner.go:211] docker network inspect addons-928332 returned with exit code 1
I0920 17:27:18.444988 536038 network_create.go:287] error running [docker network inspect addons-928332]: docker network inspect addons-928332: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-928332 not found
I0920 17:27:18.445005 536038 network_create.go:289] output of [docker network inspect addons-928332]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-928332 not found
** /stderr **
I0920 17:27:18.445090 536038 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0920 17:27:18.459807 536038 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc000015ea0}
I0920 17:27:18.459849 536038 network_create.go:124] attempt to create docker network addons-928332 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0920 17:27:18.459884 536038 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-928332 addons-928332
I0920 17:27:18.517171 536038 network_create.go:108] docker network addons-928332 192.168.49.0/24 created
I0920 17:27:18.517197 536038 kic.go:121] calculated static IP "192.168.49.2" for the "addons-928332" container
I0920 17:27:18.517249 536038 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0920 17:27:18.531710 536038 cli_runner.go:164] Run: docker volume create addons-928332 --label name.minikube.sigs.k8s.io=addons-928332 --label created_by.minikube.sigs.k8s.io=true
I0920 17:27:18.547417 536038 oci.go:103] Successfully created a docker volume addons-928332
I0920 17:27:18.547477 536038 cli_runner.go:164] Run: docker run --rm --name addons-928332-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-928332 --entrypoint /usr/bin/test -v addons-928332:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -d /var/lib
I0920 17:27:25.431089 536038 cli_runner.go:217] Completed: docker run --rm --name addons-928332-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-928332 --entrypoint /usr/bin/test -v addons-928332:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -d /var/lib: (6.883568697s)
I0920 17:27:25.431117 536038 oci.go:107] Successfully prepared a docker volume addons-928332
I0920 17:27:25.431140 536038 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0920 17:27:25.431164 536038 kic.go:194] Starting extracting preloaded images to volume ...
I0920 17:27:25.431214 536038 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19679-527899/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-928332:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -I lz4 -xf /preloaded.tar -C /extractDir
I0920 17:27:29.213809 536038 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19679-527899/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-928332:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -I lz4 -xf /preloaded.tar -C /extractDir: (3.782552941s)
I0920 17:27:29.213846 536038 kic.go:203] duration metric: took 3.782676727s to extract preloaded images to volume ...
W0920 17:27:29.213978 536038 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0920 17:27:29.214068 536038 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0920 17:27:29.263321 536038 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-928332 --name addons-928332 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-928332 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-928332 --network addons-928332 --ip 192.168.49.2 --volume addons-928332:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4
I0920 17:27:29.538307 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Running}}
I0920 17:27:29.555407 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:29.572861 536038 cli_runner.go:164] Run: docker exec addons-928332 stat /var/lib/dpkg/alternatives/iptables
I0920 17:27:29.611991 536038 oci.go:144] the created container "addons-928332" has a running status.
I0920 17:27:29.612028 536038 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa...
I0920 17:27:30.062190 536038 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0920 17:27:30.082433 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:30.098458 536038 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0920 17:27:30.098480 536038 kic_runner.go:114] Args: [docker exec --privileged addons-928332 chown docker:docker /home/docker/.ssh/authorized_keys]
I0920 17:27:30.137443 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:30.152437 536038 machine.go:93] provisionDockerMachine start ...
I0920 17:27:30.152509 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:30.169441 536038 main.go:141] libmachine: Using SSH client type: native
I0920 17:27:30.169716 536038 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33164 <nil> <nil>}
I0920 17:27:30.169736 536038 main.go:141] libmachine: About to run SSH command:
hostname
I0920 17:27:30.300234 536038 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-928332
I0920 17:27:30.300265 536038 ubuntu.go:169] provisioning hostname "addons-928332"
I0920 17:27:30.300325 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:30.316027 536038 main.go:141] libmachine: Using SSH client type: native
I0920 17:27:30.316212 536038 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33164 <nil> <nil>}
I0920 17:27:30.316227 536038 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-928332 && echo "addons-928332" | sudo tee /etc/hostname
I0920 17:27:30.458332 536038 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-928332
I0920 17:27:30.458411 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:30.473656 536038 main.go:141] libmachine: Using SSH client type: native
I0920 17:27:30.473854 536038 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33164 <nil> <nil>}
I0920 17:27:30.473871 536038 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-928332' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-928332/g' /etc/hosts;
else
echo '127.0.1.1 addons-928332' | sudo tee -a /etc/hosts;
fi
fi
I0920 17:27:30.600658 536038 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0920 17:27:30.600688 536038 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19679-527899/.minikube CaCertPath:/home/jenkins/minikube-integration/19679-527899/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19679-527899/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19679-527899/.minikube}
I0920 17:27:30.600719 536038 ubuntu.go:177] setting up certificates
I0920 17:27:30.600735 536038 provision.go:84] configureAuth start
I0920 17:27:30.600787 536038 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-928332
I0920 17:27:30.616311 536038 provision.go:143] copyHostCerts
I0920 17:27:30.616393 536038 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19679-527899/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19679-527899/.minikube/ca.pem (1082 bytes)
I0920 17:27:30.616523 536038 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19679-527899/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19679-527899/.minikube/cert.pem (1123 bytes)
I0920 17:27:30.616618 536038 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19679-527899/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19679-527899/.minikube/key.pem (1675 bytes)
I0920 17:27:30.616701 536038 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19679-527899/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19679-527899/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19679-527899/.minikube/certs/ca-key.pem org=jenkins.addons-928332 san=[127.0.0.1 192.168.49.2 addons-928332 localhost minikube]
I0920 17:27:30.711999 536038 provision.go:177] copyRemoteCerts
I0920 17:27:30.712064 536038 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0920 17:27:30.712114 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:30.727249 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:30.820930 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0920 17:27:30.841109 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I0920 17:27:30.861527 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I0920 17:27:30.881062 536038 provision.go:87] duration metric: took 280.314327ms to configureAuth
I0920 17:27:30.881099 536038 ubuntu.go:193] setting minikube options for container-runtime
I0920 17:27:30.881247 536038 config.go:182] Loaded profile config "addons-928332": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0920 17:27:30.881289 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:30.897596 536038 main.go:141] libmachine: Using SSH client type: native
I0920 17:27:30.897782 536038 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33164 <nil> <nil>}
I0920 17:27:30.897803 536038 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0920 17:27:31.025019 536038 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0920 17:27:31.025048 536038 ubuntu.go:71] root file system type: overlay
I0920 17:27:31.025222 536038 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0920 17:27:31.025292 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:31.040637 536038 main.go:141] libmachine: Using SSH client type: native
I0920 17:27:31.040815 536038 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33164 <nil> <nil>}
I0920 17:27:31.040888 536038 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0920 17:27:31.178667 536038 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0920 17:27:31.178746 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:31.194701 536038 main.go:141] libmachine: Using SSH client type: native
I0920 17:27:31.194867 536038 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x864a40] 0x867720 <nil> [] 0s} 127.0.0.1 33164 <nil> <nil>}
I0920 17:27:31.194894 536038 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0920 17:27:31.845599 536038 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-09-06 12:06:41.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-20 17:27:31.177017609 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0920 17:27:31.845630 536038 machine.go:96] duration metric: took 1.693172473s to provisionDockerMachine
I0920 17:27:31.845646 536038 client.go:171] duration metric: took 13.616876108s to LocalClient.Create
I0920 17:27:31.845667 536038 start.go:167] duration metric: took 13.616962807s to libmachine.API.Create "addons-928332"
I0920 17:27:31.845678 536038 start.go:293] postStartSetup for "addons-928332" (driver="docker")
I0920 17:27:31.845694 536038 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0920 17:27:31.845758 536038 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0920 17:27:31.845815 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:31.864413 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:31.957146 536038 ssh_runner.go:195] Run: cat /etc/os-release
I0920 17:27:31.959898 536038 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0920 17:27:31.959936 536038 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0920 17:27:31.959949 536038 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0920 17:27:31.959958 536038 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0920 17:27:31.959976 536038 filesync.go:126] Scanning /home/jenkins/minikube-integration/19679-527899/.minikube/addons for local assets ...
I0920 17:27:31.960034 536038 filesync.go:126] Scanning /home/jenkins/minikube-integration/19679-527899/.minikube/files for local assets ...
I0920 17:27:31.960065 536038 start.go:296] duration metric: took 114.376933ms for postStartSetup
I0920 17:27:31.960356 536038 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-928332
I0920 17:27:31.976041 536038 profile.go:143] Saving config to /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/config.json ...
I0920 17:27:31.976238 536038 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0920 17:27:31.976273 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:31.990998 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:32.081330 536038 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0920 17:27:32.085079 536038 start.go:128] duration metric: took 13.862018183s to createHost
I0920 17:27:32.085101 536038 start.go:83] releasing machines lock for "addons-928332", held for 13.862180904s
I0920 17:27:32.085159 536038 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-928332
I0920 17:27:32.100160 536038 ssh_runner.go:195] Run: cat /version.json
I0920 17:27:32.100210 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:32.100228 536038 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0920 17:27:32.100290 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:32.115873 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:32.117456 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:32.268030 536038 ssh_runner.go:195] Run: systemctl --version
I0920 17:27:32.271883 536038 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0920 17:27:32.275674 536038 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0920 17:27:32.297050 536038 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0920 17:27:32.297099 536038 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0920 17:27:32.319704 536038 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0920 17:27:32.319723 536038 start.go:495] detecting cgroup driver to use...
I0920 17:27:32.319748 536038 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0920 17:27:32.320247 536038 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0920 17:27:32.334658 536038 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0920 17:27:32.342628 536038 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0920 17:27:32.350793 536038 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0920 17:27:32.350841 536038 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0920 17:27:32.359237 536038 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0920 17:27:32.367021 536038 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0920 17:27:32.375345 536038 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0920 17:27:32.383069 536038 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0920 17:27:32.390961 536038 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0920 17:27:32.399170 536038 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0920 17:27:32.407196 536038 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0920 17:27:32.415010 536038 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0920 17:27:32.421714 536038 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0920 17:27:32.428344 536038 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 17:27:32.501863 536038 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0920 17:27:32.591053 536038 start.go:495] detecting cgroup driver to use...
I0920 17:27:32.591104 536038 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0920 17:27:32.591157 536038 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0920 17:27:32.601656 536038 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0920 17:27:32.601723 536038 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0920 17:27:32.611687 536038 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0920 17:27:32.626214 536038 ssh_runner.go:195] Run: which cri-dockerd
I0920 17:27:32.629287 536038 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0920 17:27:32.637573 536038 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0920 17:27:32.653959 536038 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0920 17:27:32.732893 536038 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0920 17:27:32.828114 536038 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0920 17:27:32.828271 536038 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0920 17:27:32.844601 536038 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 17:27:32.926482 536038 ssh_runner.go:195] Run: sudo systemctl restart docker
I0920 17:27:33.169137 536038 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0920 17:27:33.179907 536038 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0920 17:27:33.189682 536038 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0920 17:27:33.264531 536038 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0920 17:27:33.332958 536038 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 17:27:33.403632 536038 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0920 17:27:33.414940 536038 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0920 17:27:33.424835 536038 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 17:27:33.503365 536038 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0920 17:27:33.561068 536038 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0920 17:27:33.561160 536038 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0920 17:27:33.565037 536038 start.go:563] Will wait 60s for crictl version
I0920 17:27:33.565087 536038 ssh_runner.go:195] Run: which crictl
I0920 17:27:33.567962 536038 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0920 17:27:33.599427 536038 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.2.1
RuntimeApiVersion: v1
I0920 17:27:33.599493 536038 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0920 17:27:33.621423 536038 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0920 17:27:33.645666 536038 out.go:235] * Preparing Kubernetes v1.31.1 on Docker 27.2.1 ...
I0920 17:27:33.645754 536038 cli_runner.go:164] Run: docker network inspect addons-928332 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0920 17:27:33.661217 536038 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0920 17:27:33.664213 536038 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0920 17:27:33.673845 536038 kubeadm.go:883] updating cluster {Name:addons-928332 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-928332 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0920 17:27:33.673958 536038 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0920 17:27:33.673994 536038 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0920 17:27:33.690864 536038 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0920 17:27:33.690881 536038 docker.go:615] Images already preloaded, skipping extraction
I0920 17:27:33.690936 536038 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0920 17:27:33.707837 536038 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0920 17:27:33.707865 536038 cache_images.go:84] Images are preloaded, skipping loading
I0920 17:27:33.707876 536038 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.1 docker true true} ...
I0920 17:27:33.708020 536038 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-928332 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.1 ClusterName:addons-928332 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0920 17:27:33.708081 536038 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0920 17:27:33.749588 536038 cni.go:84] Creating CNI manager for ""
I0920 17:27:33.749613 536038 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0920 17:27:33.749622 536038 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0920 17:27:33.749639 536038 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-928332 NodeName:addons-928332 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0920 17:27:33.749775 536038 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-928332"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0920 17:27:33.749824 536038 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.1
I0920 17:27:33.757454 536038 binaries.go:44] Found k8s binaries, skipping transfer
I0920 17:27:33.757503 536038 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0920 17:27:33.764713 536038 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0920 17:27:33.779951 536038 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0920 17:27:33.794818 536038 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0920 17:27:33.809252 536038 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0920 17:27:33.812036 536038 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0920 17:27:33.821116 536038 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 17:27:33.890329 536038 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0920 17:27:33.902229 536038 certs.go:68] Setting up /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332 for IP: 192.168.49.2
I0920 17:27:33.902249 536038 certs.go:194] generating shared ca certs ...
I0920 17:27:33.902267 536038 certs.go:226] acquiring lock for ca certs: {Name:mk972fb1c4f0efdc6f550ae9a18552d0d72aa076 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:33.902400 536038 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19679-527899/.minikube/ca.key
I0920 17:27:33.965834 536038 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19679-527899/.minikube/ca.crt ...
I0920 17:27:33.965860 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/ca.crt: {Name:mkcd367569aeac77baceef380f209dc18aae8ee5 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:33.966009 536038 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19679-527899/.minikube/ca.key ...
I0920 17:27:33.966020 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/ca.key: {Name:mk2c187da0d4769605a66d27a5d874ca2b916305 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:33.966097 536038 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19679-527899/.minikube/proxy-client-ca.key
I0920 17:27:34.450971 536038 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19679-527899/.minikube/proxy-client-ca.crt ...
I0920 17:27:34.450996 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/proxy-client-ca.crt: {Name:mk3aa6c4fb139c57ec78785d34a25c6d308c339e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:34.451135 536038 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19679-527899/.minikube/proxy-client-ca.key ...
I0920 17:27:34.451145 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/proxy-client-ca.key: {Name:mk62b379060e9a888a9423bc6ef593b69461efa0 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:34.451209 536038 certs.go:256] generating profile certs ...
I0920 17:27:34.451264 536038 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/client.key
I0920 17:27:34.451278 536038 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/client.crt with IP's: []
I0920 17:27:34.545224 536038 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/client.crt ...
I0920 17:27:34.545245 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/client.crt: {Name:mkeff237f61650e5063c8cc4aefa3377c319e0a7 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:34.545365 536038 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/client.key ...
I0920 17:27:34.545375 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/client.key: {Name:mkc62c8d9e350cfd715218ff5c84d1897c81914a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:34.545452 536038 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.key.3c52f95e
I0920 17:27:34.545470 536038 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.crt.3c52f95e with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0920 17:27:34.626983 536038 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.crt.3c52f95e ...
I0920 17:27:34.627001 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.crt.3c52f95e: {Name:mkc0d267a610a1d39c13b072771413c08157bb95 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:34.627103 536038 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.key.3c52f95e ...
I0920 17:27:34.627115 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.key.3c52f95e: {Name:mk56925f2600fa38c2247339216c9a9ab4866966 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:34.627171 536038 certs.go:381] copying /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.crt.3c52f95e -> /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.crt
I0920 17:27:34.627238 536038 certs.go:385] copying /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.key.3c52f95e -> /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.key
I0920 17:27:34.627284 536038 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/proxy-client.key
I0920 17:27:34.627300 536038 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/proxy-client.crt with IP's: []
I0920 17:27:34.899626 536038 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/proxy-client.crt ...
I0920 17:27:34.899663 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/proxy-client.crt: {Name:mk1a9c26b567a8878f06b18be5165e4d1119c1f8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:34.899807 536038 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/proxy-client.key ...
I0920 17:27:34.899824 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/proxy-client.key: {Name:mk433768a8d477612c9c4226d6373fc3072a49b5 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:34.899990 536038 certs.go:484] found cert: /home/jenkins/minikube-integration/19679-527899/.minikube/certs/ca-key.pem (1675 bytes)
I0920 17:27:34.900025 536038 certs.go:484] found cert: /home/jenkins/minikube-integration/19679-527899/.minikube/certs/ca.pem (1082 bytes)
I0920 17:27:34.900048 536038 certs.go:484] found cert: /home/jenkins/minikube-integration/19679-527899/.minikube/certs/cert.pem (1123 bytes)
I0920 17:27:34.900071 536038 certs.go:484] found cert: /home/jenkins/minikube-integration/19679-527899/.minikube/certs/key.pem (1675 bytes)
I0920 17:27:34.900632 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0920 17:27:34.922123 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0920 17:27:34.941702 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0920 17:27:34.961211 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0920 17:27:34.980219 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0920 17:27:34.999554 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0920 17:27:35.019437 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0920 17:27:35.039159 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/profiles/addons-928332/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0920 17:27:35.058718 536038 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19679-527899/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0920 17:27:35.077950 536038 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0920 17:27:35.092191 536038 ssh_runner.go:195] Run: openssl version
I0920 17:27:35.096717 536038 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0920 17:27:35.104532 536038 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0920 17:27:35.107429 536038 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 20 17:27 /usr/share/ca-certificates/minikubeCA.pem
I0920 17:27:35.107479 536038 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0920 17:27:35.113231 536038 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0920 17:27:35.120676 536038 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0920 17:27:35.123336 536038 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0920 17:27:35.123374 536038 kubeadm.go:392] StartCluster: {Name:addons-928332 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-928332 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0920 17:27:35.123472 536038 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0920 17:27:35.139499 536038 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0920 17:27:35.146606 536038 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0920 17:27:35.153891 536038 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0920 17:27:35.153929 536038 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0920 17:27:35.160861 536038 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0920 17:27:35.160874 536038 kubeadm.go:157] found existing configuration files:
I0920 17:27:35.160903 536038 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0920 17:27:35.168011 536038 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0920 17:27:35.168048 536038 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0920 17:27:35.174804 536038 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0920 17:27:35.181938 536038 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0920 17:27:35.181985 536038 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0920 17:27:35.188684 536038 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0920 17:27:35.196219 536038 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0920 17:27:35.196259 536038 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0920 17:27:35.203214 536038 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0920 17:27:35.210185 536038 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0920 17:27:35.210224 536038 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0920 17:27:35.217034 536038 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0920 17:27:35.250656 536038 kubeadm.go:310] W0920 17:27:35.250127 1918 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0920 17:27:35.251490 536038 kubeadm.go:310] W0920 17:27:35.250898 1918 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0920 17:27:35.270871 536038 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1069-gcp\n", err: exit status 1
I0920 17:27:35.321709 536038 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0920 17:27:44.447353 536038 kubeadm.go:310] [init] Using Kubernetes version: v1.31.1
I0920 17:27:44.447414 536038 kubeadm.go:310] [preflight] Running pre-flight checks
I0920 17:27:44.447486 536038 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0920 17:27:44.447533 536038 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1069-gcp[0m
I0920 17:27:44.447571 536038 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0920 17:27:44.447611 536038 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0920 17:27:44.447654 536038 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0920 17:27:44.447700 536038 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0920 17:27:44.447742 536038 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0920 17:27:44.447828 536038 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0920 17:27:44.447937 536038 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0920 17:27:44.448003 536038 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0920 17:27:44.448074 536038 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0920 17:27:44.448122 536038 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0920 17:27:44.448205 536038 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0920 17:27:44.448349 536038 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0920 17:27:44.448475 536038 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0920 17:27:44.448542 536038 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0920 17:27:44.449958 536038 out.go:235] - Generating certificates and keys ...
I0920 17:27:44.450047 536038 kubeadm.go:310] [certs] Using existing ca certificate authority
I0920 17:27:44.450108 536038 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0920 17:27:44.450180 536038 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0920 17:27:44.450270 536038 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0920 17:27:44.450361 536038 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0920 17:27:44.450412 536038 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0920 17:27:44.450482 536038 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0920 17:27:44.450585 536038 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-928332 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0920 17:27:44.450630 536038 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0920 17:27:44.450738 536038 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-928332 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0920 17:27:44.450801 536038 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0920 17:27:44.450854 536038 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0920 17:27:44.450905 536038 kubeadm.go:310] [certs] Generating "sa" key and public key
I0920 17:27:44.450955 536038 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0920 17:27:44.451003 536038 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0920 17:27:44.451056 536038 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0920 17:27:44.451117 536038 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0920 17:27:44.451173 536038 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0920 17:27:44.451227 536038 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0920 17:27:44.451345 536038 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0920 17:27:44.451455 536038 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0920 17:27:44.452820 536038 out.go:235] - Booting up control plane ...
I0920 17:27:44.452950 536038 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0920 17:27:44.453078 536038 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0920 17:27:44.453188 536038 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0920 17:27:44.453328 536038 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0920 17:27:44.453453 536038 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0920 17:27:44.453507 536038 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0920 17:27:44.453632 536038 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0920 17:27:44.453718 536038 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0920 17:27:44.453767 536038 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 501.389648ms
I0920 17:27:44.453825 536038 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0920 17:27:44.453887 536038 kubeadm.go:310] [api-check] The API server is healthy after 5.002094331s
I0920 17:27:44.453997 536038 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0920 17:27:44.454121 536038 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0920 17:27:44.454178 536038 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0920 17:27:44.454357 536038 kubeadm.go:310] [mark-control-plane] Marking the node addons-928332 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0920 17:27:44.454436 536038 kubeadm.go:310] [bootstrap-token] Using token: e324ql.01ug7kp9u0jkve8y
I0920 17:27:44.455692 536038 out.go:235] - Configuring RBAC rules ...
I0920 17:27:44.455804 536038 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0920 17:27:44.455893 536038 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0920 17:27:44.456044 536038 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0920 17:27:44.456220 536038 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0920 17:27:44.456344 536038 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0920 17:27:44.456455 536038 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0920 17:27:44.456597 536038 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0920 17:27:44.456667 536038 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0920 17:27:44.456738 536038 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0920 17:27:44.456751 536038 kubeadm.go:310]
I0920 17:27:44.456839 536038 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0920 17:27:44.456846 536038 kubeadm.go:310]
I0920 17:27:44.456962 536038 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0920 17:27:44.456978 536038 kubeadm.go:310]
I0920 17:27:44.457014 536038 kubeadm.go:310] mkdir -p $HOME/.kube
I0920 17:27:44.457070 536038 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0920 17:27:44.457122 536038 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0920 17:27:44.457128 536038 kubeadm.go:310]
I0920 17:27:44.457176 536038 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0920 17:27:44.457182 536038 kubeadm.go:310]
I0920 17:27:44.457221 536038 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0920 17:27:44.457227 536038 kubeadm.go:310]
I0920 17:27:44.457299 536038 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0920 17:27:44.457434 536038 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0920 17:27:44.457526 536038 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0920 17:27:44.457535 536038 kubeadm.go:310]
I0920 17:27:44.457650 536038 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0920 17:27:44.457725 536038 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0920 17:27:44.457733 536038 kubeadm.go:310]
I0920 17:27:44.457800 536038 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token e324ql.01ug7kp9u0jkve8y \
I0920 17:27:44.457894 536038 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:90937d0c5ce288e86a7d74e3dbc8001b1c44d6b0d366b61f439e7cad2879ea6d \
I0920 17:27:44.457917 536038 kubeadm.go:310] --control-plane
I0920 17:27:44.457923 536038 kubeadm.go:310]
I0920 17:27:44.458010 536038 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0920 17:27:44.458023 536038 kubeadm.go:310]
I0920 17:27:44.458108 536038 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token e324ql.01ug7kp9u0jkve8y \
I0920 17:27:44.458210 536038 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:90937d0c5ce288e86a7d74e3dbc8001b1c44d6b0d366b61f439e7cad2879ea6d
I0920 17:27:44.458232 536038 cni.go:84] Creating CNI manager for ""
I0920 17:27:44.458254 536038 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0920 17:27:44.459815 536038 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0920 17:27:44.460868 536038 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0920 17:27:44.469015 536038 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0920 17:27:44.484506 536038 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0920 17:27:44.484562 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:44.484619 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-928332 minikube.k8s.io/updated_at=2024_09_20T17_27_44_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=35d0eeb96573bd708dfd5c070da844e6f0fad78a minikube.k8s.io/name=addons-928332 minikube.k8s.io/primary=true
I0920 17:27:44.556451 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:44.556573 536038 ops.go:34] apiserver oom_adj: -16
I0920 17:27:45.056785 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:45.556982 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:46.057194 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:46.556502 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:47.056878 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:47.556542 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:48.056787 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:48.557220 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:49.057149 536038 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0920 17:27:49.128221 536038 kubeadm.go:1113] duration metric: took 4.643707294s to wait for elevateKubeSystemPrivileges
I0920 17:27:49.128266 536038 kubeadm.go:394] duration metric: took 14.004894068s to StartCluster
I0920 17:27:49.128288 536038 settings.go:142] acquiring lock: {Name:mk85d8494b58d997c5e916d660b1eac71667663a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:49.128385 536038 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19679-527899/kubeconfig
I0920 17:27:49.128826 536038 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19679-527899/kubeconfig: {Name:mk6e28e8cbc2c8294ce12f0a9c34970facfd2529 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0920 17:27:49.129059 536038 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0920 17:27:49.129075 536038 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0920 17:27:49.129152 536038 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0920 17:27:49.129274 536038 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-928332"
I0920 17:27:49.129290 536038 addons.go:69] Setting inspektor-gadget=true in profile "addons-928332"
I0920 17:27:49.129308 536038 addons.go:234] Setting addon inspektor-gadget=true in "addons-928332"
I0920 17:27:49.129305 536038 addons.go:69] Setting gcp-auth=true in profile "addons-928332"
I0920 17:27:49.129317 536038 addons.go:69] Setting storage-provisioner=true in profile "addons-928332"
I0920 17:27:49.129327 536038 addons.go:69] Setting ingress=true in profile "addons-928332"
I0920 17:27:49.129333 536038 addons.go:69] Setting default-storageclass=true in profile "addons-928332"
I0920 17:27:49.129342 536038 addons.go:234] Setting addon ingress=true in "addons-928332"
I0920 17:27:49.129346 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.129350 536038 addons.go:69] Setting metrics-server=true in profile "addons-928332"
I0920 17:27:49.129351 536038 addons.go:69] Setting volcano=true in profile "addons-928332"
I0920 17:27:49.129357 536038 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-928332"
I0920 17:27:49.129361 536038 addons.go:234] Setting addon metrics-server=true in "addons-928332"
I0920 17:27:49.129364 536038 addons.go:69] Setting volumesnapshots=true in profile "addons-928332"
I0920 17:27:49.129371 536038 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-928332"
I0920 17:27:49.129375 536038 addons.go:234] Setting addon volumesnapshots=true in "addons-928332"
I0920 17:27:49.129382 536038 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-928332"
I0920 17:27:49.129410 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.129417 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.129427 536038 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-928332"
I0920 17:27:49.129440 536038 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-928332"
I0920 17:27:49.129458 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.129328 536038 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-928332"
I0920 17:27:49.129548 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.129743 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.129769 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.129914 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.129932 536038 addons.go:69] Setting registry=true in profile "addons-928332"
I0920 17:27:49.129953 536038 addons.go:234] Setting addon registry=true in "addons-928332"
I0920 17:27:49.129957 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.129973 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.129351 536038 mustload.go:65] Loading cluster: addons-928332
I0920 17:27:49.130353 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.129338 536038 addons.go:69] Setting cloud-spanner=true in profile "addons-928332"
I0920 17:27:49.130362 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.130378 536038 addons.go:234] Setting addon cloud-spanner=true in "addons-928332"
I0920 17:27:49.130406 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.129287 536038 addons.go:69] Setting ingress-dns=true in profile "addons-928332"
I0920 17:27:49.130438 536038 addons.go:234] Setting addon ingress-dns=true in "addons-928332"
I0920 17:27:49.130440 536038 config.go:182] Loaded profile config "addons-928332": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0920 17:27:49.130473 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.130663 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.130885 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.129338 536038 addons.go:234] Setting addon storage-provisioner=true in "addons-928332"
I0920 17:27:49.131356 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.129417 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.129364 536038 addons.go:234] Setting addon volcano=true in "addons-928332"
I0920 17:27:49.131566 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.130939 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.131832 536038 out.go:177] * Verifying Kubernetes components...
I0920 17:27:49.129278 536038 addons.go:69] Setting yakd=true in profile "addons-928332"
I0920 17:27:49.131905 536038 addons.go:234] Setting addon yakd=true in "addons-928332"
I0920 17:27:49.131968 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.129923 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.129314 536038 config.go:182] Loaded profile config "addons-928332": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0920 17:27:49.129918 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.133187 536038 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0920 17:27:49.157990 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.158069 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.158408 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.158464 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.164220 536038 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0920 17:27:49.165817 536038 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-928332"
I0920 17:27:49.165869 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.166329 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.167482 536038 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0920 17:27:49.168447 536038 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0920 17:27:49.169503 536038 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0920 17:27:49.170895 536038 addons.go:234] Setting addon default-storageclass=true in "addons-928332"
I0920 17:27:49.170938 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.171396 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:49.171627 536038 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0920 17:27:49.172708 536038 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0920 17:27:49.173682 536038 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0920 17:27:49.175121 536038 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0920 17:27:49.176156 536038 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0920 17:27:49.176182 536038 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0920 17:27:49.176244 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.188961 536038 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0920 17:27:49.188971 536038 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0920 17:27:49.190511 536038 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0920 17:27:49.190539 536038 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0920 17:27:49.190618 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.190816 536038 out.go:177] - Using image docker.io/registry:2.8.3
I0920 17:27:49.191875 536038 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0920 17:27:49.191907 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0920 17:27:49.191960 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.195747 536038 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0920 17:27:49.196783 536038 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0920 17:27:49.197835 536038 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0920 17:27:49.199029 536038 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0920 17:27:49.199055 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0920 17:27:49.199120 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.210317 536038 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.24
I0920 17:27:49.212045 536038 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0920 17:27:49.212288 536038 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0920 17:27:49.212307 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0920 17:27:49.212594 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.213250 536038 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0920 17:27:49.213327 536038 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0920 17:27:49.213525 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.216072 536038 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0920 17:27:49.216251 536038 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0920 17:27:49.217550 536038 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0920 17:27:49.217569 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0920 17:27:49.217622 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.219756 536038 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0920 17:27:49.219778 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0920 17:27:49.219829 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.220334 536038 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0920 17:27:49.221660 536038 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0920 17:27:49.221678 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0920 17:27:49.221728 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.221933 536038 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0920 17:27:49.222972 536038 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0920 17:27:49.222993 536038 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0920 17:27:49.223049 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.234580 536038 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0920 17:27:49.235712 536038 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0920 17:27:49.235736 536038 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0920 17:27:49.235809 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.246074 536038 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0920 17:27:49.246092 536038 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0920 17:27:49.246149 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.264241 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.264616 536038 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0920 17:27:49.264831 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:49.266478 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.266725 536038 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0920 17:27:49.268280 536038 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0920 17:27:49.269478 536038 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0920 17:27:49.270061 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.270704 536038 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0920 17:27:49.270728 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0920 17:27:49.270778 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.271636 536038 out.go:177] - Using image docker.io/busybox:stable
I0920 17:27:49.272927 536038 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0920 17:27:49.272950 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0920 17:27:49.272994 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:49.276595 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.278187 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.280380 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.283384 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.286780 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.305655 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.306045 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.306100 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.323650 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.323648 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.326672 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:49.402906 536038 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0920 17:27:49.403853 536038 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0920 17:27:49.691633 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0920 17:27:49.691697 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0920 17:27:49.889900 536038 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0920 17:27:49.890000 536038 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0920 17:27:49.898122 536038 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0920 17:27:49.898158 536038 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0920 17:27:49.900570 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0920 17:27:49.907278 536038 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0920 17:27:49.907302 536038 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0920 17:27:49.907971 536038 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0920 17:27:49.908005 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0920 17:27:49.910690 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0920 17:27:49.997962 536038 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0920 17:27:49.997989 536038 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0920 17:27:49.998384 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0920 17:27:50.001272 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0920 17:27:50.090744 536038 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0920 17:27:50.090777 536038 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0920 17:27:50.096291 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0920 17:27:50.193017 536038 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0920 17:27:50.193119 536038 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0920 17:27:50.205080 536038 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0920 17:27:50.205159 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0920 17:27:50.208678 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0920 17:27:50.210748 536038 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0920 17:27:50.210815 536038 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0920 17:27:50.496831 536038 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0920 17:27:50.496921 536038 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0920 17:27:50.498681 536038 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0920 17:27:50.498745 536038 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0920 17:27:50.602050 536038 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0920 17:27:50.602145 536038 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0920 17:27:50.603644 536038 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0920 17:27:50.603712 536038 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0920 17:27:50.691508 536038 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0920 17:27:50.691609 536038 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0920 17:27:50.897154 536038 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0920 17:27:50.897246 536038 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0920 17:27:50.907916 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0920 17:27:50.910301 536038 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0920 17:27:50.910388 536038 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0920 17:27:50.991465 536038 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.587578595s)
I0920 17:27:50.992536 536038 node_ready.go:35] waiting up to 6m0s for node "addons-928332" to be "Ready" ...
I0920 17:27:50.992791 536038 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.589848391s)
I0920 17:27:50.992817 536038 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0920 17:27:50.996947 536038 node_ready.go:49] node "addons-928332" has status "Ready":"True"
I0920 17:27:50.997011 536038 node_ready.go:38] duration metric: took 4.442386ms for node "addons-928332" to be "Ready" ...
I0920 17:27:50.997053 536038 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0920 17:27:51.006371 536038 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace to be "Ready" ...
I0920 17:27:51.190405 536038 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0920 17:27:51.190492 536038 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0920 17:27:51.208869 536038 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0920 17:27:51.208896 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0920 17:27:51.209843 536038 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0920 17:27:51.209880 536038 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0920 17:27:51.298307 536038 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0920 17:27:51.298408 536038 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0920 17:27:51.309722 536038 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0920 17:27:51.309805 536038 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0920 17:27:51.493802 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0920 17:27:51.499073 536038 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-928332" context rescaled to 1 replicas
I0920 17:27:51.600932 536038 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0920 17:27:51.600961 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0920 17:27:51.608093 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0920 17:27:51.992682 536038 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0920 17:27:51.992713 536038 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0920 17:27:52.091856 536038 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0920 17:27:52.091972 536038 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0920 17:27:52.204767 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0920 17:27:52.493252 536038 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0920 17:27:52.493346 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0920 17:27:52.809351 536038 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0920 17:27:52.809444 536038 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0920 17:27:52.911170 536038 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0920 17:27:52.911199 536038 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0920 17:27:53.093238 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:27:53.192178 536038 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0920 17:27:53.192255 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0920 17:27:53.306531 536038 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0920 17:27:53.306612 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0920 17:27:53.608652 536038 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0920 17:27:53.608748 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0920 17:27:53.697002 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0920 17:27:53.895209 536038 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0920 17:27:53.895253 536038 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0920 17:27:54.193775 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0920 17:27:55.105067 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:27:56.297318 536038 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0920 17:27:56.297452 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:56.322578 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:57.098248 536038 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0920 17:27:57.306055 536038 addons.go:234] Setting addon gcp-auth=true in "addons-928332"
I0920 17:27:57.306186 536038 host.go:66] Checking if "addons-928332" exists ...
I0920 17:27:57.306851 536038 cli_runner.go:164] Run: docker container inspect addons-928332 --format={{.State.Status}}
I0920 17:27:57.326670 536038 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0920 17:27:57.326725 536038 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-928332
I0920 17:27:57.341581 536038 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:33164 SSHKeyPath:/home/jenkins/minikube-integration/19679-527899/.minikube/machines/addons-928332/id_rsa Username:docker}
I0920 17:27:57.515952 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:27:58.005225 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (8.313483137s)
I0920 17:27:58.005350 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (8.104752276s)
I0920 17:27:58.005694 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (8.094950665s)
I0920 17:27:58.005767 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (8.007335488s)
I0920 17:27:58.005823 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (8.004500264s)
I0920 17:27:58.005857 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (7.909485551s)
I0920 17:27:58.006427 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (8.314751585s)
I0920 17:27:58.006451 536038 addons.go:475] Verifying addon ingress=true in "addons-928332"
I0920 17:27:58.008007 536038 out.go:177] * Verifying ingress addon...
I0920 17:27:58.011152 536038 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
W0920 17:27:58.091839 536038 out.go:270] ! Enabling 'default-storageclass' returned an error: running callbacks: [Error making standard the default storage class: Error while marking storage class local-path as non-default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
I0920 17:27:58.094446 536038 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0920 17:27:58.094520 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:27:58.597426 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:27:59.095585 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:27:59.519151 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:27:59.519397 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:00.106427 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:00.592796 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:01.091869 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:01.111186 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (10.902399s)
I0920 17:28:01.111271 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (10.203260784s)
I0920 17:28:01.111311 536038 addons.go:475] Verifying addon registry=true in "addons-928332"
I0920 17:28:01.111670 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (9.617819692s)
I0920 17:28:01.112000 536038 addons.go:475] Verifying addon metrics-server=true in "addons-928332"
I0920 17:28:01.111937 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (8.907076925s)
W0920 17:28:01.112075 536038 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0920 17:28:01.112090 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (7.415010417s)
I0920 17:28:01.112159 536038 retry.go:31] will retry after 170.116862ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0920 17:28:01.111972 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (9.503787855s)
I0920 17:28:01.112841 536038 out.go:177] * Verifying registry addon...
I0920 17:28:01.113661 536038 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-928332 service yakd-dashboard -n yakd-dashboard
I0920 17:28:01.115554 536038 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0920 17:28:01.192849 536038 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0920 17:28:01.192919 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:01.283300 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0920 17:28:01.591324 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:01.593674 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:01.691759 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:02.015750 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:02.119029 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:02.219250 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (8.025406792s)
I0920 17:28:02.219487 536038 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-928332"
I0920 17:28:02.219415 536038 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (4.892720904s)
I0920 17:28:02.220746 536038 out.go:177] * Verifying csi-hostpath-driver addon...
I0920 17:28:02.220759 536038 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0920 17:28:02.221931 536038 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0920 17:28:02.222715 536038 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0920 17:28:02.222941 536038 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0920 17:28:02.222958 536038 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0920 17:28:02.294115 536038 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0920 17:28:02.294149 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:02.314040 536038 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0920 17:28:02.314104 536038 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0920 17:28:02.396121 536038 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0920 17:28:02.396231 536038 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0920 17:28:02.498022 536038 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0920 17:28:02.524348 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:02.618889 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:02.795294 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:03.016322 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:03.193165 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:03.294622 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:03.515442 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:03.619092 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:03.793257 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:03.812921 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.529564856s)
I0920 17:28:03.930172 536038 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.432091722s)
I0920 17:28:03.931563 536038 addons.go:475] Verifying addon gcp-auth=true in "addons-928332"
I0920 17:28:03.932759 536038 out.go:177] * Verifying gcp-auth addon...
I0920 17:28:03.934373 536038 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0920 17:28:03.993523 536038 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0920 17:28:04.011352 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:04.014280 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:04.119550 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:04.227559 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:04.515152 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:04.619817 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:04.727816 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:05.038321 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:05.138568 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:05.226373 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:05.515001 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:05.619094 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:05.728077 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:06.012057 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:06.014579 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:06.118644 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:06.227202 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:06.514786 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:06.619169 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:06.727135 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:07.040018 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:07.140170 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:07.226615 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:07.513792 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:07.618745 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:07.784565 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:08.013968 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:08.118608 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:08.226778 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:08.511563 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:08.513859 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:08.618785 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:08.726686 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:09.018363 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:09.118585 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:09.227985 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:09.515241 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:09.619400 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:09.727355 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:10.014760 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:10.119097 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:10.227678 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:10.514527 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:10.618318 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:10.726540 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:11.010760 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:11.038691 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:11.118314 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:11.227529 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:11.515572 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:11.618945 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:11.727536 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:12.014912 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:12.118858 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:12.227168 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:12.514621 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:12.619034 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:12.727858 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:13.011900 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:13.014436 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:13.119568 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:13.227220 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:13.514819 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:13.619387 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:13.728328 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:14.014881 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:14.118634 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:14.227584 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:14.514895 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:14.619276 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:14.728729 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:15.012612 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:15.015015 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:15.119566 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:15.226920 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:15.514141 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:15.619860 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:15.727473 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:16.014496 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:16.118859 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:16.227983 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:16.515360 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:16.619981 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:16.728328 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:17.013563 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:17.014385 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:17.119430 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:17.226599 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:17.514654 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:17.619563 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:17.727247 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:18.013933 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:18.118485 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:18.226470 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:18.515773 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:18.619165 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:18.726849 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:19.039279 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:19.139918 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:19.227533 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:19.512398 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:19.514566 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:19.618564 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:19.727041 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:20.014289 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:20.120092 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:20.227472 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:20.514533 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:20.619023 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:20.727558 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:21.014378 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:21.119705 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:21.227215 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:21.514024 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:21.619211 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0920 17:28:21.727535 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:22.011556 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:22.039236 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:22.119467 536038 kapi.go:107] duration metric: took 21.00390395s to wait for kubernetes.io/minikube-addons=registry ...
I0920 17:28:22.226831 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:22.514777 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:22.727496 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:23.039473 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:23.228031 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:23.515251 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:23.727820 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:24.014093 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:24.227528 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:24.512529 536038 pod_ready.go:103] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"False"
I0920 17:28:24.514691 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:24.727448 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:25.014730 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:25.227395 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:25.514477 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:25.727366 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:26.011994 536038 pod_ready.go:93] pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace has status "Ready":"True"
I0920 17:28:26.012022 536038 pod_ready.go:82] duration metric: took 35.005589017s for pod "coredns-7c65d6cfc9-gm5h9" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.012035 536038 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-kmjc6" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.013606 536038 pod_ready.go:98] error getting pod "coredns-7c65d6cfc9-kmjc6" in "kube-system" namespace (skipping!): pods "coredns-7c65d6cfc9-kmjc6" not found
I0920 17:28:26.013632 536038 pod_ready.go:82] duration metric: took 1.588211ms for pod "coredns-7c65d6cfc9-kmjc6" in "kube-system" namespace to be "Ready" ...
E0920 17:28:26.013644 536038 pod_ready.go:67] WaitExtra: waitPodCondition: error getting pod "coredns-7c65d6cfc9-kmjc6" in "kube-system" namespace (skipping!): pods "coredns-7c65d6cfc9-kmjc6" not found
I0920 17:28:26.013654 536038 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-928332" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.018087 536038 pod_ready.go:93] pod "etcd-addons-928332" in "kube-system" namespace has status "Ready":"True"
I0920 17:28:26.018106 536038 pod_ready.go:82] duration metric: took 4.443624ms for pod "etcd-addons-928332" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.018115 536038 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-928332" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.022640 536038 pod_ready.go:93] pod "kube-apiserver-addons-928332" in "kube-system" namespace has status "Ready":"True"
I0920 17:28:26.022663 536038 pod_ready.go:82] duration metric: took 4.540903ms for pod "kube-apiserver-addons-928332" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.022674 536038 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-928332" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.027003 536038 pod_ready.go:93] pod "kube-controller-manager-addons-928332" in "kube-system" namespace has status "Ready":"True"
I0920 17:28:26.027022 536038 pod_ready.go:82] duration metric: took 4.340071ms for pod "kube-controller-manager-addons-928332" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.027033 536038 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-mqpkw" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.091401 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:26.210122 536038 pod_ready.go:93] pod "kube-proxy-mqpkw" in "kube-system" namespace has status "Ready":"True"
I0920 17:28:26.210143 536038 pod_ready.go:82] duration metric: took 183.103644ms for pod "kube-proxy-mqpkw" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.210154 536038 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-928332" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.226385 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:26.514912 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:26.610270 536038 pod_ready.go:93] pod "kube-scheduler-addons-928332" in "kube-system" namespace has status "Ready":"True"
I0920 17:28:26.610297 536038 pod_ready.go:82] duration metric: took 400.135838ms for pod "kube-scheduler-addons-928332" in "kube-system" namespace to be "Ready" ...
I0920 17:28:26.610304 536038 pod_ready.go:39] duration metric: took 35.613222807s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0920 17:28:26.610328 536038 api_server.go:52] waiting for apiserver process to appear ...
I0920 17:28:26.610384 536038 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0920 17:28:26.624267 536038 api_server.go:72] duration metric: took 37.495156444s to wait for apiserver process to appear ...
I0920 17:28:26.624287 536038 api_server.go:88] waiting for apiserver healthz status ...
I0920 17:28:26.624308 536038 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0920 17:28:26.627691 536038 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0920 17:28:26.628467 536038 api_server.go:141] control plane version: v1.31.1
I0920 17:28:26.628495 536038 api_server.go:131] duration metric: took 4.20167ms to wait for apiserver health ...
I0920 17:28:26.628504 536038 system_pods.go:43] waiting for kube-system pods to appear ...
I0920 17:28:26.727748 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:26.816661 536038 system_pods.go:59] 17 kube-system pods found
I0920 17:28:26.816692 536038 system_pods.go:61] "coredns-7c65d6cfc9-gm5h9" [663c229a-031b-49d4-9e29-7ba464e69235] Running
I0920 17:28:26.816701 536038 system_pods.go:61] "csi-hostpath-attacher-0" [40bdbc72-5d98-4d7c-8d86-36f880b6ddd2] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0920 17:28:26.816708 536038 system_pods.go:61] "csi-hostpath-resizer-0" [c7992406-52ce-4ba2-933a-ce0a0e357af1] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0920 17:28:26.816716 536038 system_pods.go:61] "csi-hostpathplugin-qxjls" [93ec10a6-20d1-476b-a960-732e0f966089] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0920 17:28:26.816720 536038 system_pods.go:61] "etcd-addons-928332" [d45eb170-182d-40d1-ab14-81a23ff18af5] Running
I0920 17:28:26.816724 536038 system_pods.go:61] "kube-apiserver-addons-928332" [5c74f7ea-cd77-42b2-a319-f15e74f071c2] Running
I0920 17:28:26.816728 536038 system_pods.go:61] "kube-controller-manager-addons-928332" [3a698e91-0cb8-4976-8179-cc45f29f2480] Running
I0920 17:28:26.816732 536038 system_pods.go:61] "kube-ingress-dns-minikube" [41a0d111-9601-4042-8b60-128a898a9fb7] Running
I0920 17:28:26.816737 536038 system_pods.go:61] "kube-proxy-mqpkw" [06a3ff6e-e3c2-4a11-8722-95d5feb51fcc] Running
I0920 17:28:26.816741 536038 system_pods.go:61] "kube-scheduler-addons-928332" [4352a83a-e6f3-4e78-b2e6-57e2ea7214b2] Running
I0920 17:28:26.816750 536038 system_pods.go:61] "metrics-server-84c5f94fbc-gnz59" [fedfa55d-e1ba-414f-a821-d5855d940137] Running
I0920 17:28:26.816754 536038 system_pods.go:61] "nvidia-device-plugin-daemonset-lwwgw" [608331f5-f175-4ab8-a481-4fe212f174dd] Running
I0920 17:28:26.816757 536038 system_pods.go:61] "registry-66c9cd494c-vmjl4" [83533b4a-57e0-46f9-8a17-5cbf2c880aea] Running
I0920 17:28:26.816760 536038 system_pods.go:61] "registry-proxy-n8nl5" [60333ce0-1ed5-49fd-bf03-d4ba227481ce] Running
I0920 17:28:26.816766 536038 system_pods.go:61] "snapshot-controller-56fcc65765-46scd" [1b2ff531-13fa-4c65-bda6-2b20fcf72917] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 17:28:26.816773 536038 system_pods.go:61] "snapshot-controller-56fcc65765-rkl99" [57681adc-ace8-476c-8aa6-9a8a4efc045f] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 17:28:26.816780 536038 system_pods.go:61] "storage-provisioner" [048e3d59-4e1d-40cf-88a2-cf606f2c796d] Running
I0920 17:28:26.816787 536038 system_pods.go:74] duration metric: took 188.273147ms to wait for pod list to return data ...
I0920 17:28:26.816796 536038 default_sa.go:34] waiting for default service account to be created ...
I0920 17:28:27.010473 536038 default_sa.go:45] found service account: "default"
I0920 17:28:27.010562 536038 default_sa.go:55] duration metric: took 193.70568ms for default service account to be created ...
I0920 17:28:27.010581 536038 system_pods.go:116] waiting for k8s-apps to be running ...
I0920 17:28:27.014975 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:27.217015 536038 system_pods.go:86] 17 kube-system pods found
I0920 17:28:27.217051 536038 system_pods.go:89] "coredns-7c65d6cfc9-gm5h9" [663c229a-031b-49d4-9e29-7ba464e69235] Running
I0920 17:28:27.217072 536038 system_pods.go:89] "csi-hostpath-attacher-0" [40bdbc72-5d98-4d7c-8d86-36f880b6ddd2] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0920 17:28:27.217083 536038 system_pods.go:89] "csi-hostpath-resizer-0" [c7992406-52ce-4ba2-933a-ce0a0e357af1] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0920 17:28:27.217101 536038 system_pods.go:89] "csi-hostpathplugin-qxjls" [93ec10a6-20d1-476b-a960-732e0f966089] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0920 17:28:27.217108 536038 system_pods.go:89] "etcd-addons-928332" [d45eb170-182d-40d1-ab14-81a23ff18af5] Running
I0920 17:28:27.217114 536038 system_pods.go:89] "kube-apiserver-addons-928332" [5c74f7ea-cd77-42b2-a319-f15e74f071c2] Running
I0920 17:28:27.217121 536038 system_pods.go:89] "kube-controller-manager-addons-928332" [3a698e91-0cb8-4976-8179-cc45f29f2480] Running
I0920 17:28:27.217130 536038 system_pods.go:89] "kube-ingress-dns-minikube" [41a0d111-9601-4042-8b60-128a898a9fb7] Running
I0920 17:28:27.217135 536038 system_pods.go:89] "kube-proxy-mqpkw" [06a3ff6e-e3c2-4a11-8722-95d5feb51fcc] Running
I0920 17:28:27.217141 536038 system_pods.go:89] "kube-scheduler-addons-928332" [4352a83a-e6f3-4e78-b2e6-57e2ea7214b2] Running
I0920 17:28:27.217147 536038 system_pods.go:89] "metrics-server-84c5f94fbc-gnz59" [fedfa55d-e1ba-414f-a821-d5855d940137] Running
I0920 17:28:27.217158 536038 system_pods.go:89] "nvidia-device-plugin-daemonset-lwwgw" [608331f5-f175-4ab8-a481-4fe212f174dd] Running
I0920 17:28:27.217165 536038 system_pods.go:89] "registry-66c9cd494c-vmjl4" [83533b4a-57e0-46f9-8a17-5cbf2c880aea] Running
I0920 17:28:27.217174 536038 system_pods.go:89] "registry-proxy-n8nl5" [60333ce0-1ed5-49fd-bf03-d4ba227481ce] Running
I0920 17:28:27.217185 536038 system_pods.go:89] "snapshot-controller-56fcc65765-46scd" [1b2ff531-13fa-4c65-bda6-2b20fcf72917] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 17:28:27.217194 536038 system_pods.go:89] "snapshot-controller-56fcc65765-rkl99" [57681adc-ace8-476c-8aa6-9a8a4efc045f] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0920 17:28:27.217207 536038 system_pods.go:89] "storage-provisioner" [048e3d59-4e1d-40cf-88a2-cf606f2c796d] Running
I0920 17:28:27.217215 536038 system_pods.go:126] duration metric: took 206.623809ms to wait for k8s-apps to be running ...
I0920 17:28:27.217224 536038 system_svc.go:44] waiting for kubelet service to be running ....
I0920 17:28:27.217274 536038 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0920 17:28:27.227525 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:27.231738 536038 system_svc.go:56] duration metric: took 14.501682ms WaitForService to wait for kubelet
I0920 17:28:27.231770 536038 kubeadm.go:582] duration metric: took 38.102661064s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0920 17:28:27.231797 536038 node_conditions.go:102] verifying NodePressure condition ...
I0920 17:28:27.411393 536038 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0920 17:28:27.411427 536038 node_conditions.go:123] node cpu capacity is 8
I0920 17:28:27.411444 536038 node_conditions.go:105] duration metric: took 179.640065ms to run NodePressure ...
I0920 17:28:27.411459 536038 start.go:241] waiting for startup goroutines ...
I0920 17:28:27.514926 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:27.727542 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:28.015187 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:28.227869 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:28.515431 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:28.762636 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:29.015950 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:29.226788 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:29.515072 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:29.727854 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:30.039822 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:30.292691 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:30.515122 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:30.727602 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:31.015431 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:31.227877 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:31.514853 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:31.727831 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:32.015772 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:32.226307 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:32.515395 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:32.728363 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:33.015168 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:33.227448 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:33.514591 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:33.727365 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:34.015123 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:34.227700 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:34.514504 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:34.727431 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:35.015688 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:35.228063 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:35.539093 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:35.727953 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:36.015966 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:36.227438 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:36.515396 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:36.727920 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:37.039304 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:37.227066 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:37.515560 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:37.728159 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:38.064650 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:38.226648 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:38.515148 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:38.727480 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:39.038511 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:39.226345 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:39.514880 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:39.727077 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:40.015940 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:40.227450 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:40.515748 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:40.727300 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:41.015172 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:41.228599 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:41.515493 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:41.727701 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:42.015134 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:42.227250 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:42.515157 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:42.727334 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:43.014564 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:43.227745 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:43.515001 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:43.774502 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:44.039297 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:44.227636 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:44.515605 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:44.728081 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:45.038715 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:45.227018 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:45.514739 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:45.727949 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0920 17:28:46.015196 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:46.226448 536038 kapi.go:107] duration metric: took 44.003731745s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0920 17:28:46.514891 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:47.014771 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:47.519318 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:48.015306 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:48.515047 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:49.014195 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:49.513827 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:50.014966 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:50.513960 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:51.015114 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:51.515252 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:52.014683 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:52.514148 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:53.014786 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:53.538340 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:54.014439 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:54.515273 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:55.014490 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:55.514013 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:56.013966 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:56.514488 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:57.014719 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:57.513932 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:58.014441 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:58.514504 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:59.015636 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:28:59.514832 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:00.014863 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:00.514576 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:01.014711 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:01.514740 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:02.014078 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:02.521863 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:03.015618 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:03.515846 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:04.015706 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:04.516070 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:05.015569 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:05.515391 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:06.015441 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:06.514932 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:07.038894 536038 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0920 17:29:07.514075 536038 kapi.go:107] duration metric: took 1m9.502921155s to wait for app.kubernetes.io/name=ingress-nginx ...
I0920 17:29:27.437074 536038 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0920 17:29:27.437100 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:27.937442 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:28.436764 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:28.938464 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:29.437518 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:29.937333 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:30.438063 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:30.937737 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:31.437421 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:31.937230 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:32.437921 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:32.937870 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:33.438118 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:33.938066 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:34.438136 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:34.938119 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:35.437963 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:35.937585 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:36.437341 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:36.937654 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:37.437276 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:37.937951 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:38.437787 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:38.937689 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:39.437591 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:39.937802 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:40.437339 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:40.937956 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:41.437689 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:41.937708 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:42.437275 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:42.938199 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:43.437855 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:43.937799 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:44.437076 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:44.937888 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:45.437736 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:45.937831 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:46.438025 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:46.937986 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:47.437628 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:47.937115 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:48.437634 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:48.937849 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:49.437757 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:49.937569 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:50.437181 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:50.938001 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:51.438028 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:51.937866 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:52.437859 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:52.937873 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:53.437723 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:53.937656 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:54.437279 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:54.938165 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:55.437927 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:55.937786 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:56.437672 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:56.938064 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:57.437712 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:57.937687 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:58.437726 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:58.937763 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:59.437663 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:29:59.937187 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:00.437676 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:00.937608 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:01.437367 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:01.937935 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:02.437612 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:02.937435 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:03.437120 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:03.937602 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:04.437213 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:04.938063 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:05.437775 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:05.937651 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:06.437594 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:06.937821 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:07.438582 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:07.938238 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:08.438085 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:08.937853 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:09.437980 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:09.938364 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:10.437015 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:10.937764 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:11.437441 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:11.937271 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:12.438019 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:12.938516 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:13.437583 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:13.937004 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:14.437847 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:14.938134 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:15.437906 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:15.937802 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:16.437972 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:16.954859 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:17.437361 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:17.937317 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:18.437549 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:18.937688 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:19.437489 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:19.937536 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:20.437199 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:20.937888 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:21.437939 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:21.938299 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:22.436968 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:22.937763 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:23.437407 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:23.937994 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:24.437806 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:24.938461 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:25.437277 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:25.937020 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:26.437871 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:26.938198 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:27.437905 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:27.937540 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:28.437001 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:28.937163 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:29.437298 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:29.938360 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:30.438032 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:30.938248 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:31.438549 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:31.938108 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:32.438312 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:32.938375 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:33.437998 536038 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0920 17:30:33.937406 536038 kapi.go:107] duration metric: took 2m30.003014955s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0920 17:30:33.938732 536038 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-928332 cluster.
I0920 17:30:33.939947 536038 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0920 17:30:33.940997 536038 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0920 17:30:33.942196 536038 out.go:177] * Enabled addons: cloud-spanner, storage-provisioner, nvidia-device-plugin, ingress-dns, storage-provisioner-rancher, volcano, metrics-server, inspektor-gadget, yakd, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0920 17:30:33.943140 536038 addons.go:510] duration metric: took 2m44.813991054s for enable addons: enabled=[cloud-spanner storage-provisioner nvidia-device-plugin ingress-dns storage-provisioner-rancher volcano metrics-server inspektor-gadget yakd volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0920 17:30:33.943181 536038 start.go:246] waiting for cluster config update ...
I0920 17:30:33.943202 536038 start.go:255] writing updated cluster config ...
I0920 17:30:33.943480 536038 ssh_runner.go:195] Run: rm -f paused
I0920 17:30:33.993940 536038 start.go:600] kubectl: 1.31.1, cluster: 1.31.1 (minor skew: 0)
I0920 17:30:33.996165 536038 out.go:177] * Done! kubectl is now configured to use "addons-928332" cluster and "default" namespace by default
==> Docker <==
Sep 20 17:40:03 addons-928332 dockerd[1337]: time="2024-09-20T17:40:03.104260476Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=da3369f2e6450880f8901b61710e2eabeb285111a76cd0bc856e45d34ecde865
Sep 20 17:40:03 addons-928332 dockerd[1337]: time="2024-09-20T17:40:03.154979822Z" level=info msg="ignoring event" container=da3369f2e6450880f8901b61710e2eabeb285111a76cd0bc856e45d34ecde865 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:03 addons-928332 dockerd[1337]: time="2024-09-20T17:40:03.278518030Z" level=info msg="ignoring event" container=8fc6310770fe644dc85d50ee7b9c0b5572e5ea3c7369ca52709e80cce14a31f4 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:06 addons-928332 dockerd[1337]: time="2024-09-20T17:40:06.782592963Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 20 17:40:06 addons-928332 dockerd[1337]: time="2024-09-20T17:40:06.784149051Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 20 17:40:16 addons-928332 cri-dockerd[1601]: time="2024-09-20T17:40:16Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/b9d0babc451ac055bce8a5df55483b2be62e77909eb2f62f7e4cd735cc5f0c41/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local us-east4-a.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 20 17:40:16 addons-928332 cri-dockerd[1601]: time="2024-09-20T17:40:16Z" level=info msg="Stop pulling image docker.io/nginx:latest: Status: Image is up to date for nginx:latest"
Sep 20 17:40:23 addons-928332 dockerd[1337]: time="2024-09-20T17:40:23.125291410Z" level=info msg="ignoring event" container=5cf7c39d199073fe0bd6f185e4ff9693e200a0aba01060dd5264eb0c2604fd7a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:23 addons-928332 dockerd[1337]: time="2024-09-20T17:40:23.239239791Z" level=info msg="ignoring event" container=b9d0babc451ac055bce8a5df55483b2be62e77909eb2f62f7e4cd735cc5f0c41 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:24 addons-928332 dockerd[1337]: time="2024-09-20T17:40:24.380616564Z" level=info msg="ignoring event" container=f4859a2934b8de1562e8066c26e7adf9c93fa1fdad99943278f0192121f42223 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:24 addons-928332 dockerd[1337]: time="2024-09-20T17:40:24.895843296Z" level=info msg="ignoring event" container=9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:24 addons-928332 dockerd[1337]: time="2024-09-20T17:40:24.900214072Z" level=info msg="ignoring event" container=99f54f41d98f1ef49dc859b20622f5cc2c6173072ff4f60c1d163e37c8f081b2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:24 addons-928332 dockerd[1337]: time="2024-09-20T17:40:24.908009004Z" level=info msg="ignoring event" container=3614ca971a11ca6f67917b679b33f20fc544129e6698416ea352dea0d1361310 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:24 addons-928332 dockerd[1337]: time="2024-09-20T17:40:24.909528831Z" level=info msg="ignoring event" container=193d1cf0d851e9bf1fe3ad8ad3aacd5efece6d549316cd0566039eb5f2473327 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:24 addons-928332 dockerd[1337]: time="2024-09-20T17:40:24.910719793Z" level=info msg="ignoring event" container=d642f81a8d20dcddc25b41abf406f8dbe940cae00ece1f90bbeb053421580a85 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:24 addons-928332 dockerd[1337]: time="2024-09-20T17:40:24.912493577Z" level=info msg="ignoring event" container=71e609c2b2de10278ba71fbf4ba151cf777716b96f687f2905edf5929415b02c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:24 addons-928332 dockerd[1337]: time="2024-09-20T17:40:24.914089038Z" level=info msg="ignoring event" container=8a4a8c6fa3fdcb55d0b2bbbf78066edc81c86324053bba6c66ec0ecfb282b971 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:25 addons-928332 dockerd[1337]: time="2024-09-20T17:40:25.012098392Z" level=info msg="ignoring event" container=b4468f573f0e521733d7daf2efbb57af9e8b057051da25bdd474f4382fd3c772 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:25 addons-928332 dockerd[1337]: time="2024-09-20T17:40:25.291359770Z" level=info msg="ignoring event" container=4ba2ecb545575dfe8dfb1801dc39e68c1de70c86ff1a0dfe5b4d9693d8b4c7a8 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:25 addons-928332 dockerd[1337]: time="2024-09-20T17:40:25.316816206Z" level=info msg="ignoring event" container=274ff6507095c697dc246f439538800dbcf8cad8c574c1c44601b1cfa53f7483 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:25 addons-928332 dockerd[1337]: time="2024-09-20T17:40:25.393745819Z" level=info msg="ignoring event" container=1bdf76d11d2716e5bf0f06ee064305f67224da3d789837f864dc7fcecf53e163 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:25 addons-928332 dockerd[1337]: time="2024-09-20T17:40:25.398403055Z" level=info msg="ignoring event" container=52c73397b53c45a5ebd429bb433b5a72088d7d37861d045ca58821c927241486 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:25 addons-928332 dockerd[1337]: time="2024-09-20T17:40:25.412737131Z" level=info msg="ignoring event" container=a624e099f56c14926a24cc495c207a2812c92d90afac1e868502d355f1b1fa81 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:25 addons-928332 dockerd[1337]: time="2024-09-20T17:40:25.598697098Z" level=info msg="ignoring event" container=57b903d542e4e8f6c18b44cd88d9b66e75f5bce0e7c7feb815b5bdd41c484a4a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 20 17:40:25 addons-928332 dockerd[1337]: time="2024-09-20T17:40:25.719111506Z" level=info msg="ignoring event" container=8b59128f55417dcb254f9a0632a53da857ac29065a7684190a62a94f39683226 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
adb77004bb285 kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 28 seconds ago Running hello-world-app 0 30a7ff0d08ba2 hello-world-app-55bf9c44b4-l4wmj
0539599bb4a39 nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf 37 seconds ago Running nginx 0 ee084dd4156b5 nginx
e927ab3a39d4a gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 244955f3e09a5 gcp-auth-89d5ffd79-pd8ft
e65790c5b0059 registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 11 minutes ago Running volume-snapshot-controller 0 497aafe45474f snapshot-controller-56fcc65765-46scd
2a0e519948f2f registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 11 minutes ago Running volume-snapshot-controller 0 d8ac3f537c685 snapshot-controller-56fcc65765-rkl99
7635c727999b0 ce263a8653f9c 11 minutes ago Exited patch 1 8c19c9e5da99f ingress-nginx-admission-patch-w5lx6
d2aeaf1e1c1fa registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 44b93bd06edb5 ingress-nginx-admission-create-ddr42
40aced01d29cb 6e38f40d628db 12 minutes ago Running storage-provisioner 0 25ff0c651f4b3 storage-provisioner
d1bf3cc1b0df5 c69fa2e9cbf5f 12 minutes ago Running coredns 0 5edfd6c5521e0 coredns-7c65d6cfc9-gm5h9
eb4cd368494e4 60c005f310ff3 12 minutes ago Running kube-proxy 0 592e4d6514120 kube-proxy-mqpkw
89cfc9d895d00 175ffd71cce3d 12 minutes ago Running kube-controller-manager 0 4d478da377e8f kube-controller-manager-addons-928332
cb06ff32b3c20 6bab7719df100 12 minutes ago Running kube-apiserver 0 3683fc3e44ac4 kube-apiserver-addons-928332
8c9ef0872c3ff 2e96e5913fc06 12 minutes ago Running etcd 0 42c4c34a25adf etcd-addons-928332
46d2bb774ff16 9aa1fad941575 12 minutes ago Running kube-scheduler 0 4e137deadf961 kube-scheduler-addons-928332
==> coredns [d1bf3cc1b0df] <==
[INFO] 10.244.0.21:42178 - 39440 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005210399s
[INFO] 10.244.0.21:51947 - 27645 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005077637s
[INFO] 10.244.0.21:35287 - 54707 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.0050493s
[INFO] 10.244.0.21:59190 - 10324 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005188387s
[INFO] 10.244.0.21:36836 - 55972 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005642506s
[INFO] 10.244.0.21:47842 - 13834 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.006637032s
[INFO] 10.244.0.21:53512 - 28093 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00664343s
[INFO] 10.244.0.21:35287 - 16879 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004523532s
[INFO] 10.244.0.21:51947 - 47236 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003444223s
[INFO] 10.244.0.21:60269 - 53421 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005044963s
[INFO] 10.244.0.21:47842 - 61352 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.002660247s
[INFO] 10.244.0.21:59190 - 24865 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003900791s
[INFO] 10.244.0.21:35287 - 63430 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000087839s
[INFO] 10.244.0.21:36836 - 13604 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005343938s
[INFO] 10.244.0.21:53512 - 10359 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003673615s
[INFO] 10.244.0.21:60269 - 10446 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000085615s
[INFO] 10.244.0.21:51947 - 28097 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000072736s
[INFO] 10.244.0.21:42178 - 7138 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005346179s
[INFO] 10.244.0.21:36086 - 48486 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00555562s
[INFO] 10.244.0.21:53512 - 16517 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000075473s
[INFO] 10.244.0.21:47842 - 7640 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000129429s
[INFO] 10.244.0.21:36086 - 13837 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000059637s
[INFO] 10.244.0.21:36836 - 1340 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00006249s
[INFO] 10.244.0.21:59190 - 42383 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00007881s
[INFO] 10.244.0.21:42178 - 16101 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000140463s
==> describe nodes <==
Name: addons-928332
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-928332
kubernetes.io/os=linux
minikube.k8s.io/commit=35d0eeb96573bd708dfd5c070da844e6f0fad78a
minikube.k8s.io/name=addons-928332
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_20T17_27_44_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-928332
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Fri, 20 Sep 2024 17:27:41 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-928332
AcquireTime: <unset>
RenewTime: Fri, 20 Sep 2024 17:40:18 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Fri, 20 Sep 2024 17:40:17 +0000 Fri, 20 Sep 2024 17:27:39 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Fri, 20 Sep 2024 17:40:17 +0000 Fri, 20 Sep 2024 17:27:39 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Fri, 20 Sep 2024 17:40:17 +0000 Fri, 20 Sep 2024 17:27:39 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Fri, 20 Sep 2024 17:40:17 +0000 Fri, 20 Sep 2024 17:27:41 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-928332
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
System Info:
Machine ID: 4b79b85d3a78481aa604e7f663f20820
System UUID: f17ae1ac-a446-4070-b14a-37ac70e1ca5f
Boot ID: b374bddd-f952-49e6-aac3-b4f18d49c77c
Kernel Version: 5.15.0-1069-gcp
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.2.1
Kubelet Version: v1.31.1
Kube-Proxy Version: v1.31.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (13 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m15s
default hello-world-app-55bf9c44b4-l4wmj 0 (0%) 0 (0%) 0 (0%) 0 (0%) 29s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 39s
gcp-auth gcp-auth-89d5ffd79-pd8ft 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
kube-system coredns-7c65d6cfc9-gm5h9 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-928332 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-928332 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-928332 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-mqpkw 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-928332 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system snapshot-controller-56fcc65765-46scd 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system snapshot-controller-56fcc65765-rkl99 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-928332 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-928332 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-928332 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-928332 event: Registered Node addons-928332 in Controller
==> dmesg <==
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff c2 6c 94 6a 98 c5 08 06
[ +1.344937] IPv4: martian source 10.244.0.1 from 10.244.0.16, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 72 60 69 0e 5d 8e 08 06
[ +1.465079] IPv4: martian source 10.244.0.1 from 10.244.0.17, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff f2 6f f3 85 de dd 08 06
[ +5.286716] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 0a 34 ee e9 56 a7 08 06
[ +0.624912] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 1e a2 06 e4 55 74 08 06
[ +0.094155] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 62 18 d8 8a e2 3f 08 06
[Sep20 17:29] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 4a d0 03 65 19 5e 08 06
[Sep20 17:30] IPv4: martian source 10.244.0.1 from 10.244.0.23, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 7a 30 34 28 6c d0 08 06
[ +0.022462] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff 96 bf 86 66 14 8b 08 06
[ +25.649241] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff ce cf 2c 59 94 a4 08 06
[ +0.000467] IPv4: martian source 10.244.0.25 from 10.244.0.2, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff 0e 63 23 87 50 42 08 06
[Sep20 17:39] IPv4: martian source 10.244.0.33 from 10.244.0.21, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 4a d0 03 65 19 5e 08 06
[ +0.586563] IPv4: martian source 10.244.0.21 from 10.244.0.2, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 0e 63 23 87 50 42 08 06
==> etcd [8c9ef0872c3f] <==
{"level":"info","ts":"2024-09-20T17:27:38.734582Z","caller":"embed/etcd.go:599","msg":"serving peer traffic","address":"192.168.49.2:2380"}
{"level":"info","ts":"2024-09-20T17:27:38.789736Z","caller":"embed/etcd.go:571","msg":"cmux::serve","address":"192.168.49.2:2380"}
{"level":"info","ts":"2024-09-20T17:27:39.321566Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc is starting a new election at term 1"}
{"level":"info","ts":"2024-09-20T17:27:39.321616Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became pre-candidate at term 1"}
{"level":"info","ts":"2024-09-20T17:27:39.321643Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgPreVoteResp from aec36adc501070cc at term 1"}
{"level":"info","ts":"2024-09-20T17:27:39.321664Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became candidate at term 2"}
{"level":"info","ts":"2024-09-20T17:27:39.321673Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-20T17:27:39.321682Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-09-20T17:27:39.321692Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-20T17:27:39.322678Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-928332 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-09-20T17:27:39.322684Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-20T17:27:39.322691Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T17:27:39.322713Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-20T17:27:39.323086Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-20T17:27:39.323142Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-20T17:27:39.323787Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-20T17:27:39.323792Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-20T17:27:39.324592Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-09-20T17:27:39.324598Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-20T17:27:39.324861Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T17:27:39.324949Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T17:27:39.324975Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-20T17:37:39.397381Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1855}
{"level":"info","ts":"2024-09-20T17:37:39.419787Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1855,"took":"21.851199ms","hash":1999188028,"current-db-size-bytes":9072640,"current-db-size":"9.1 MB","current-db-size-in-use-bytes":4866048,"current-db-size-in-use":"4.9 MB"}
{"level":"info","ts":"2024-09-20T17:37:39.419825Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":1999188028,"revision":1855,"compact-revision":-1}
==> gcp-auth [e927ab3a39d4] <==
2024/09/20 17:31:11 Ready to write response ...
2024/09/20 17:31:12 Ready to marshal response ...
2024/09/20 17:31:12 Ready to write response ...
2024/09/20 17:39:14 Ready to marshal response ...
2024/09/20 17:39:14 Ready to write response ...
2024/09/20 17:39:14 Ready to marshal response ...
2024/09/20 17:39:14 Ready to write response ...
2024/09/20 17:39:14 Ready to marshal response ...
2024/09/20 17:39:14 Ready to write response ...
2024/09/20 17:39:19 Ready to marshal response ...
2024/09/20 17:39:19 Ready to write response ...
2024/09/20 17:39:19 Ready to marshal response ...
2024/09/20 17:39:19 Ready to write response ...
2024/09/20 17:39:24 Ready to marshal response ...
2024/09/20 17:39:24 Ready to write response ...
2024/09/20 17:39:27 Ready to marshal response ...
2024/09/20 17:39:27 Ready to write response ...
2024/09/20 17:39:47 Ready to marshal response ...
2024/09/20 17:39:47 Ready to write response ...
2024/09/20 17:39:49 Ready to marshal response ...
2024/09/20 17:39:49 Ready to write response ...
2024/09/20 17:39:57 Ready to marshal response ...
2024/09/20 17:39:57 Ready to write response ...
2024/09/20 17:40:15 Ready to marshal response ...
2024/09/20 17:40:15 Ready to write response ...
==> kernel <==
17:40:26 up 1:22, 0 users, load average: 1.73, 1.07, 1.56
Linux addons-928332 5.15.0-1069-gcp #77~20.04.1-Ubuntu SMP Sun Sep 1 19:39:16 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [cb06ff32b3c2] <==
I0920 17:31:02.195028 1 handler.go:286] Adding GroupVersion scheduling.volcano.sh v1beta1 to ResourceManager
I0920 17:31:02.215538 1 handler.go:286] Adding GroupVersion scheduling.volcano.sh v1beta1 to ResourceManager
I0920 17:31:02.295763 1 handler.go:286] Adding GroupVersion scheduling.volcano.sh v1beta1 to ResourceManager
I0920 17:31:02.296049 1 handler.go:286] Adding GroupVersion nodeinfo.volcano.sh v1alpha1 to ResourceManager
I0920 17:31:02.798025 1 handler.go:286] Adding GroupVersion flow.volcano.sh v1alpha1 to ResourceManager
I0920 17:31:02.905211 1 handler.go:286] Adding GroupVersion flow.volcano.sh v1alpha1 to ResourceManager
I0920 17:31:02.999499 1 handler.go:286] Adding GroupVersion flow.volcano.sh v1alpha1 to ResourceManager
W0920 17:31:03.009797 1 cacher.go:171] Terminating all watchers from cacher commands.bus.volcano.sh
W0920 17:31:03.305040 1 cacher.go:171] Terminating all watchers from cacher podgroups.scheduling.volcano.sh
W0920 17:31:03.500849 1 cacher.go:171] Terminating all watchers from cacher numatopologies.nodeinfo.volcano.sh
W0920 17:31:03.597340 1 cacher.go:171] Terminating all watchers from cacher queues.scheduling.volcano.sh
W0920 17:31:03.718260 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0920 17:31:04.000025 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0920 17:31:04.308394 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0920 17:39:14.716511 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.110.82.224"}
I0920 17:39:42.149159 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0920 17:39:43.165263 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
E0920 17:39:43.290822 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
I0920 17:39:47.586898 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0920 17:39:47.817967 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.108.204.115"}
I0920 17:39:56.841356 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0920 17:39:57.365266 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.96.12.100"}
E0920 17:39:59.556844 1 watch.go:250] "Unhandled Error" err="http2: stream closed" logger="UnhandledError"
E0920 17:40:01.326853 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"ingress-nginx\" not found]"
E0920 17:40:01.332032 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"ingress-nginx\" not found]"
==> kube-controller-manager [89cfc9d895d0] <==
I0920 17:39:59.602246 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="48.174µs"
W0920 17:39:59.930390 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 17:39:59.930439 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0920 17:40:00.056169 1 job_controller.go:568] "enqueueing job" logger="job-controller" key="ingress-nginx/ingress-nginx-admission-create" delay="0s"
I0920 17:40:00.057334 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="ingress-nginx/ingress-nginx-controller-bc57996ff" duration="9.187µs"
I0920 17:40:00.059650 1 job_controller.go:568] "enqueueing job" logger="job-controller" key="ingress-nginx/ingress-nginx-admission-patch" delay="0s"
W0920 17:40:07.164062 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 17:40:07.164102 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 17:40:08.021197 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 17:40:08.021236 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0920 17:40:10.153001 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="ingress-nginx"
W0920 17:40:13.327386 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 17:40:13.327427 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0920 17:40:15.613673 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="local-path-storage"
I0920 17:40:17.920585 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-928332"
W0920 17:40:20.128201 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 17:40:20.128241 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 17:40:20.663401 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 17:40:20.663439 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0920 17:40:20.925002 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0920 17:40:20.925040 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0920 17:40:24.691928 1 stateful_set.go:466] "StatefulSet has been deleted" logger="statefulset-controller" key="kube-system/csi-hostpath-attacher"
I0920 17:40:24.793032 1 stateful_set.go:466] "StatefulSet has been deleted" logger="statefulset-controller" key="kube-system/csi-hostpath-resizer"
I0920 17:40:25.212357 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="10.381µs"
I0920 17:40:25.509270 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-928332"
==> kube-proxy [eb4cd368494e] <==
I0920 17:27:52.009348 1 server_linux.go:66] "Using iptables proxy"
I0920 17:27:52.501497 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0920 17:27:52.501583 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0920 17:27:52.905983 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0920 17:27:52.906051 1 server_linux.go:169] "Using iptables Proxier"
I0920 17:27:52.992725 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0920 17:27:52.993077 1 server.go:483] "Version info" version="v1.31.1"
I0920 17:27:52.993097 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0920 17:27:53.002172 1 config.go:328] "Starting node config controller"
I0920 17:27:53.002207 1 shared_informer.go:313] Waiting for caches to sync for node config
I0920 17:27:53.003024 1 config.go:199] "Starting service config controller"
I0920 17:27:53.003041 1 shared_informer.go:313] Waiting for caches to sync for service config
I0920 17:27:53.003261 1 config.go:105] "Starting endpoint slice config controller"
I0920 17:27:53.003272 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0920 17:27:53.109460 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0920 17:27:53.109526 1 shared_informer.go:320] Caches are synced for service config
I0920 17:27:53.110251 1 shared_informer.go:320] Caches are synced for node config
==> kube-scheduler [46d2bb774ff1] <==
W0920 17:27:41.553623 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0920 17:27:41.553702 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0920 17:27:41.553799 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0920 17:27:41.553830 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 17:27:41.553955 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0920 17:27:41.553986 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 17:27:41.554160 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0920 17:27:41.554196 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 17:27:41.554162 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0920 17:27:41.554231 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0920 17:27:41.554353 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
W0920 17:27:41.554360 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0920 17:27:41.554382 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0920 17:27:41.554384 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 17:27:41.554658 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
W0920 17:27:41.554699 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0920 17:27:41.554687 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
E0920 17:27:41.554739 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0920 17:27:41.554701 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0920 17:27:41.554782 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 17:27:42.400046 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0920 17:27:42.400097 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0920 17:27:42.435470 1 reflector.go:561] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0920 17:27:42.435515 1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
I0920 17:27:44.752201 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.251337 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705"} err="failed to get container status \"9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705\": rpc error: code = Unknown desc = Error response from daemon: No such container: 9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.251372 2435 scope.go:117] "RemoveContainer" containerID="8a4a8c6fa3fdcb55d0b2bbbf78066edc81c86324053bba6c66ec0ecfb282b971"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.251947 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"8a4a8c6fa3fdcb55d0b2bbbf78066edc81c86324053bba6c66ec0ecfb282b971"} err="failed to get container status \"8a4a8c6fa3fdcb55d0b2bbbf78066edc81c86324053bba6c66ec0ecfb282b971\": rpc error: code = Unknown desc = Error response from daemon: No such container: 8a4a8c6fa3fdcb55d0b2bbbf78066edc81c86324053bba6c66ec0ecfb282b971"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.251976 2435 scope.go:117] "RemoveContainer" containerID="193d1cf0d851e9bf1fe3ad8ad3aacd5efece6d549316cd0566039eb5f2473327"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.252573 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"193d1cf0d851e9bf1fe3ad8ad3aacd5efece6d549316cd0566039eb5f2473327"} err="failed to get container status \"193d1cf0d851e9bf1fe3ad8ad3aacd5efece6d549316cd0566039eb5f2473327\": rpc error: code = Unknown desc = Error response from daemon: No such container: 193d1cf0d851e9bf1fe3ad8ad3aacd5efece6d549316cd0566039eb5f2473327"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.252605 2435 scope.go:117] "RemoveContainer" containerID="3614ca971a11ca6f67917b679b33f20fc544129e6698416ea352dea0d1361310"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.253141 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"3614ca971a11ca6f67917b679b33f20fc544129e6698416ea352dea0d1361310"} err="failed to get container status \"3614ca971a11ca6f67917b679b33f20fc544129e6698416ea352dea0d1361310\": rpc error: code = Unknown desc = Error response from daemon: No such container: 3614ca971a11ca6f67917b679b33f20fc544129e6698416ea352dea0d1361310"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.253177 2435 scope.go:117] "RemoveContainer" containerID="d642f81a8d20dcddc25b41abf406f8dbe940cae00ece1f90bbeb053421580a85"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.253769 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"d642f81a8d20dcddc25b41abf406f8dbe940cae00ece1f90bbeb053421580a85"} err="failed to get container status \"d642f81a8d20dcddc25b41abf406f8dbe940cae00ece1f90bbeb053421580a85\": rpc error: code = Unknown desc = Error response from daemon: No such container: d642f81a8d20dcddc25b41abf406f8dbe940cae00ece1f90bbeb053421580a85"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.253796 2435 scope.go:117] "RemoveContainer" containerID="71e609c2b2de10278ba71fbf4ba151cf777716b96f687f2905edf5929415b02c"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.254420 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"71e609c2b2de10278ba71fbf4ba151cf777716b96f687f2905edf5929415b02c"} err="failed to get container status \"71e609c2b2de10278ba71fbf4ba151cf777716b96f687f2905edf5929415b02c\": rpc error: code = Unknown desc = Error response from daemon: No such container: 71e609c2b2de10278ba71fbf4ba151cf777716b96f687f2905edf5929415b02c"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.254450 2435 scope.go:117] "RemoveContainer" containerID="9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.255011 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705"} err="failed to get container status \"9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705\": rpc error: code = Unknown desc = Error response from daemon: No such container: 9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.255032 2435 scope.go:117] "RemoveContainer" containerID="8a4a8c6fa3fdcb55d0b2bbbf78066edc81c86324053bba6c66ec0ecfb282b971"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.255563 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"8a4a8c6fa3fdcb55d0b2bbbf78066edc81c86324053bba6c66ec0ecfb282b971"} err="failed to get container status \"8a4a8c6fa3fdcb55d0b2bbbf78066edc81c86324053bba6c66ec0ecfb282b971\": rpc error: code = Unknown desc = Error response from daemon: No such container: 8a4a8c6fa3fdcb55d0b2bbbf78066edc81c86324053bba6c66ec0ecfb282b971"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.255591 2435 scope.go:117] "RemoveContainer" containerID="193d1cf0d851e9bf1fe3ad8ad3aacd5efece6d549316cd0566039eb5f2473327"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.256184 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"193d1cf0d851e9bf1fe3ad8ad3aacd5efece6d549316cd0566039eb5f2473327"} err="failed to get container status \"193d1cf0d851e9bf1fe3ad8ad3aacd5efece6d549316cd0566039eb5f2473327\": rpc error: code = Unknown desc = Error response from daemon: No such container: 193d1cf0d851e9bf1fe3ad8ad3aacd5efece6d549316cd0566039eb5f2473327"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.256217 2435 scope.go:117] "RemoveContainer" containerID="3614ca971a11ca6f67917b679b33f20fc544129e6698416ea352dea0d1361310"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.256745 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"3614ca971a11ca6f67917b679b33f20fc544129e6698416ea352dea0d1361310"} err="failed to get container status \"3614ca971a11ca6f67917b679b33f20fc544129e6698416ea352dea0d1361310\": rpc error: code = Unknown desc = Error response from daemon: No such container: 3614ca971a11ca6f67917b679b33f20fc544129e6698416ea352dea0d1361310"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.256767 2435 scope.go:117] "RemoveContainer" containerID="d642f81a8d20dcddc25b41abf406f8dbe940cae00ece1f90bbeb053421580a85"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.290007 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"d642f81a8d20dcddc25b41abf406f8dbe940cae00ece1f90bbeb053421580a85"} err="failed to get container status \"d642f81a8d20dcddc25b41abf406f8dbe940cae00ece1f90bbeb053421580a85\": rpc error: code = Unknown desc = Error response from daemon: No such container: d642f81a8d20dcddc25b41abf406f8dbe940cae00ece1f90bbeb053421580a85"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.290030 2435 scope.go:117] "RemoveContainer" containerID="71e609c2b2de10278ba71fbf4ba151cf777716b96f687f2905edf5929415b02c"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.290674 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"71e609c2b2de10278ba71fbf4ba151cf777716b96f687f2905edf5929415b02c"} err="failed to get container status \"71e609c2b2de10278ba71fbf4ba151cf777716b96f687f2905edf5929415b02c\": rpc error: code = Unknown desc = Error response from daemon: No such container: 71e609c2b2de10278ba71fbf4ba151cf777716b96f687f2905edf5929415b02c"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.290711 2435 scope.go:117] "RemoveContainer" containerID="9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705"
Sep 20 17:40:26 addons-928332 kubelet[2435]: I0920 17:40:26.291348 2435 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705"} err="failed to get container status \"9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705\": rpc error: code = Unknown desc = Error response from daemon: No such container: 9bd8ac8e0e2dc36a7ca1787bdc401b6f864e63aa4ec661706c0d701485e75705"
==> storage-provisioner [40aced01d29c] <==
I0920 17:27:56.692392 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0920 17:27:56.703933 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0920 17:27:56.704017 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0920 17:27:56.791024 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0920 17:27:56.791239 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-928332_3755d3a6-32fd-42d9-bedf-16be71d03050!
I0920 17:27:56.792970 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"bb1bf246-015e-4ebd-be90-a77a31435cfa", APIVersion:"v1", ResourceVersion:"584", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-928332_3755d3a6-32fd-42d9-bedf-16be71d03050 became leader
I0920 17:27:56.893507 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-928332_3755d3a6-32fd-42d9-bedf-16be71d03050!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-928332 -n addons-928332
helpers_test.go:261: (dbg) Run: kubectl --context addons-928332 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-928332 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-928332 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-928332/192.168.49.2
Start Time: Fri, 20 Sep 2024 17:31:11 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.27
IPs:
IP: 10.244.0.27
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-r25s9 (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-r25s9:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m16s default-scheduler Successfully assigned default/busybox to addons-928332
Normal Pulling 7m50s (x4 over 9m15s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m50s (x4 over 9m15s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m50s (x4 over 9m15s) kubelet Error: ErrImagePull
Warning Failed 7m24s (x6 over 9m15s) kubelet Error: ImagePullBackOff
Normal BackOff 4m4s (x21 over 9m15s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (73.20s)