=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:332: registry stabilized in 1.868571ms
addons_test.go:334: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-zlfgb" [cb8eada4-03cf-4da1-8b47-5bce3b3d3226] Running
addons_test.go:334: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.003056763s
addons_test.go:337: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-jb2hn" [eccdebcf-344c-4fa8-9409-061fa7c8d2ab] Running
addons_test.go:337: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.003681687s
addons_test.go:342: (dbg) Run: kubectl --context addons-382533 delete po -l run=registry-test --now
addons_test.go:347: (dbg) Run: kubectl --context addons-382533 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:347: (dbg) Non-zero exit: kubectl --context addons-382533 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.078260586s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:349: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-382533 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:353: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:361: (dbg) Run: out/minikube-linux-amd64 -p addons-382533 ip
2024/09/10 17:43:19 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:390: (dbg) Run: out/minikube-linux-amd64 -p addons-382533 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-382533
helpers_test.go:235: (dbg) docker inspect addons-382533:
-- stdout --
[
{
"Id": "b72c8c40d158eb022758ababb4b88c5bb9a08bba618b9be4b96f873a71a19fdb",
"Created": "2024-09-10T17:30:16.447531145Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 14749,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-10T17:30:16.576229144Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:81bbf527fa1222fda707ad6c719d13a0d8bd3ffc934d975c9155a6ffac7840c6",
"ResolvConfPath": "/var/lib/docker/containers/b72c8c40d158eb022758ababb4b88c5bb9a08bba618b9be4b96f873a71a19fdb/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/b72c8c40d158eb022758ababb4b88c5bb9a08bba618b9be4b96f873a71a19fdb/hostname",
"HostsPath": "/var/lib/docker/containers/b72c8c40d158eb022758ababb4b88c5bb9a08bba618b9be4b96f873a71a19fdb/hosts",
"LogPath": "/var/lib/docker/containers/b72c8c40d158eb022758ababb4b88c5bb9a08bba618b9be4b96f873a71a19fdb/b72c8c40d158eb022758ababb4b88c5bb9a08bba618b9be4b96f873a71a19fdb-json.log",
"Name": "/addons-382533",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-382533:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-382533",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/1fad06c74591679e4335b994db031caacf8e91ba04b3bf15ab5578199ea5107f-init/diff:/var/lib/docker/overlay2/660816bf0fa4a262508d36bd0ad60b151b10cf98d92fe6b1a4362eabd9004836/diff",
"MergedDir": "/var/lib/docker/overlay2/1fad06c74591679e4335b994db031caacf8e91ba04b3bf15ab5578199ea5107f/merged",
"UpperDir": "/var/lib/docker/overlay2/1fad06c74591679e4335b994db031caacf8e91ba04b3bf15ab5578199ea5107f/diff",
"WorkDir": "/var/lib/docker/overlay2/1fad06c74591679e4335b994db031caacf8e91ba04b3bf15ab5578199ea5107f/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-382533",
"Source": "/var/lib/docker/volumes/addons-382533/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-382533",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-382533",
"name.minikube.sigs.k8s.io": "addons-382533",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "6816f9f44082774da204acbf16fef4fcd80b1f988f02081812f7122f849d71d8",
"SandboxKey": "/var/run/docker/netns/6816f9f44082",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-382533": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "2f9d28c9a6fabdac2759f31f4b9a710c6556bd8e9599c8955ca99078366f83e7",
"EndpointID": "b9054300943e3d4663f58a0db772360f8e00bae5e8522d3826247185afa2a82a",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-382533",
"b72c8c40d158"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-382533 -n addons-382533
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-382533 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-docker-182088 | download-docker-182088 | jenkins | v1.34.0 | 10 Sep 24 17:29 UTC | 10 Sep 24 17:29 UTC |
| start | --download-only -p | binary-mirror-851154 | jenkins | v1.34.0 | 10 Sep 24 17:29 UTC | |
| | binary-mirror-851154 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:46439 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-851154 | binary-mirror-851154 | jenkins | v1.34.0 | 10 Sep 24 17:29 UTC | 10 Sep 24 17:29 UTC |
| addons | enable dashboard -p | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:29 UTC | |
| | addons-382533 | | | | | |
| addons | disable dashboard -p | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:29 UTC | |
| | addons-382533 | | | | | |
| start | -p addons-382533 --wait=true | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:29 UTC | 10 Sep 24 17:33 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | addons-382533 addons disable | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:33 UTC | 10 Sep 24 17:34 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | addons-382533 | | | | | |
| addons | addons-382533 addons disable | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-382533 addons disable | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | -p addons-382533 | | | | | |
| addons | disable cloud-spanner -p | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | addons-382533 | | | | | |
| addons | enable headlamp | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | -p addons-382533 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-382533 ssh cat | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | /opt/local-path-provisioner/pvc-b9fbf517-0e38-4d92-b138-26e3220b8867_default_test-pvc/file1 | | | | | |
| addons | addons-382533 addons disable | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-382533 addons disable | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-382533 addons | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-382533 addons | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-382533 ssh curl -s | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-382533 ip | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| addons | addons-382533 addons disable | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:42 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-382533 addons disable | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:43 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| addons | addons-382533 addons | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:42 UTC | 10 Sep 24 17:43 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-382533 ip | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:43 UTC | 10 Sep 24 17:43 UTC |
| addons | addons-382533 addons disable | addons-382533 | jenkins | v1.34.0 | 10 Sep 24 17:43 UTC | 10 Sep 24 17:43 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/10 17:29:55
Running on machine: ubuntu-20-agent-10
Binary: Built with gc go1.22.5 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0910 17:29:55.046515 14014 out.go:345] Setting OutFile to fd 1 ...
I0910 17:29:55.046637 14014 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0910 17:29:55.046648 14014 out.go:358] Setting ErrFile to fd 2...
I0910 17:29:55.046654 14014 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0910 17:29:55.046863 14014 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19598-5816/.minikube/bin
I0910 17:29:55.047457 14014 out.go:352] Setting JSON to false
I0910 17:29:55.048249 14014 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-10","uptime":740,"bootTime":1725988655,"procs":172,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1067-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0910 17:29:55.048313 14014 start.go:139] virtualization: kvm guest
I0910 17:29:55.050483 14014 out.go:177] * [addons-382533] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0910 17:29:55.051698 14014 out.go:177] - MINIKUBE_LOCATION=19598
I0910 17:29:55.051706 14014 notify.go:220] Checking for updates...
I0910 17:29:55.053123 14014 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0910 17:29:55.054418 14014 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19598-5816/kubeconfig
I0910 17:29:55.055623 14014 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19598-5816/.minikube
I0910 17:29:55.056753 14014 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0910 17:29:55.058001 14014 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0910 17:29:55.059334 14014 driver.go:394] Setting default libvirt URI to qemu:///system
I0910 17:29:55.079639 14014 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
I0910 17:29:55.079729 14014 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0910 17:29:55.127915 14014 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-10 17:29:55.119268275 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1067-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-10 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:472731909fa34bd7bc9c087e4c27943f9835f111 Expected:472731909fa34bd7bc9c087e4c27943f9835f111} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0910 17:29:55.128018 14014 docker.go:318] overlay module found
I0910 17:29:55.145011 14014 out.go:177] * Using the docker driver based on user configuration
I0910 17:29:55.166409 14014 start.go:297] selected driver: docker
I0910 17:29:55.166434 14014 start.go:901] validating driver "docker" against <nil>
I0910 17:29:55.166446 14014 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0910 17:29:55.167488 14014 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0910 17:29:55.211923 14014 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-10 17:29:55.20332822 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1067-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86_
64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-10 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:472731909fa34bd7bc9c087e4c27943f9835f111 Expected:472731909fa34bd7bc9c087e4c27943f9835f111} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErr
ors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0910 17:29:55.212116 14014 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0910 17:29:55.212329 14014 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0910 17:29:55.213872 14014 out.go:177] * Using Docker driver with root privileges
I0910 17:29:55.215295 14014 cni.go:84] Creating CNI manager for ""
I0910 17:29:55.215317 14014 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0910 17:29:55.215330 14014 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0910 17:29:55.215392 14014 start.go:340] cluster config:
{Name:addons-382533 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-382533 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0910 17:29:55.216522 14014 out.go:177] * Starting "addons-382533" primary control-plane node in "addons-382533" cluster
I0910 17:29:55.217609 14014 cache.go:121] Beginning downloading kic base image for docker with docker
I0910 17:29:55.218915 14014 out.go:177] * Pulling base image v0.0.45-1725963390-19606 ...
I0910 17:29:55.220045 14014 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0910 17:29:55.220082 14014 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19598-5816/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4
I0910 17:29:55.220081 14014 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 in local docker daemon
I0910 17:29:55.220089 14014 cache.go:56] Caching tarball of preloaded images
I0910 17:29:55.220165 14014 preload.go:172] Found /home/jenkins/minikube-integration/19598-5816/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0910 17:29:55.220178 14014 cache.go:59] Finished verifying existence of preloaded tar for v1.31.0 on docker
I0910 17:29:55.220510 14014 profile.go:143] Saving config to /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/config.json ...
I0910 17:29:55.220531 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/config.json: {Name:mk3ca10f077f9d9af62a6ec9e29a9fed4565caa9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:29:55.235245 14014 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 to local cache
I0910 17:29:55.235355 14014 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 in local cache directory
I0910 17:29:55.235375 14014 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 in local cache directory, skipping pull
I0910 17:29:55.235381 14014 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 exists in cache, skipping pull
I0910 17:29:55.235393 14014 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 as a tarball
I0910 17:29:55.235398 14014 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 from local cache
I0910 17:30:07.334249 14014 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 from cached tarball
I0910 17:30:07.334299 14014 cache.go:194] Successfully downloaded all kic artifacts
I0910 17:30:07.334346 14014 start.go:360] acquireMachinesLock for addons-382533: {Name:mkb90550fbf9001475ea0a472cc64ee93239cd69 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0910 17:30:07.334456 14014 start.go:364] duration metric: took 86.529µs to acquireMachinesLock for "addons-382533"
I0910 17:30:07.334484 14014 start.go:93] Provisioning new machine with config: &{Name:addons-382533 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-382533 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0910 17:30:07.334572 14014 start.go:125] createHost starting for "" (driver="docker")
I0910 17:30:07.336392 14014 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0910 17:30:07.336628 14014 start.go:159] libmachine.API.Create for "addons-382533" (driver="docker")
I0910 17:30:07.336665 14014 client.go:168] LocalClient.Create starting
I0910 17:30:07.336754 14014 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19598-5816/.minikube/certs/ca.pem
I0910 17:30:07.501460 14014 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19598-5816/.minikube/certs/cert.pem
I0910 17:30:07.594561 14014 cli_runner.go:164] Run: docker network inspect addons-382533 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0910 17:30:07.609820 14014 cli_runner.go:211] docker network inspect addons-382533 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0910 17:30:07.609894 14014 network_create.go:284] running [docker network inspect addons-382533] to gather additional debugging logs...
I0910 17:30:07.609915 14014 cli_runner.go:164] Run: docker network inspect addons-382533
W0910 17:30:07.624724 14014 cli_runner.go:211] docker network inspect addons-382533 returned with exit code 1
I0910 17:30:07.624760 14014 network_create.go:287] error running [docker network inspect addons-382533]: docker network inspect addons-382533: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-382533 not found
I0910 17:30:07.624786 14014 network_create.go:289] output of [docker network inspect addons-382533]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-382533 not found
** /stderr **
I0910 17:30:07.624878 14014 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0910 17:30:07.640375 14014 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001c45f70}
I0910 17:30:07.640425 14014 network_create.go:124] attempt to create docker network addons-382533 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0910 17:30:07.640466 14014 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-382533 addons-382533
I0910 17:30:07.698556 14014 network_create.go:108] docker network addons-382533 192.168.49.0/24 created
I0910 17:30:07.698582 14014 kic.go:121] calculated static IP "192.168.49.2" for the "addons-382533" container
I0910 17:30:07.698633 14014 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0910 17:30:07.713188 14014 cli_runner.go:164] Run: docker volume create addons-382533 --label name.minikube.sigs.k8s.io=addons-382533 --label created_by.minikube.sigs.k8s.io=true
I0910 17:30:07.730083 14014 oci.go:103] Successfully created a docker volume addons-382533
I0910 17:30:07.730155 14014 cli_runner.go:164] Run: docker run --rm --name addons-382533-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-382533 --entrypoint /usr/bin/test -v addons-382533:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 -d /var/lib
I0910 17:30:12.496410 14014 cli_runner.go:217] Completed: docker run --rm --name addons-382533-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-382533 --entrypoint /usr/bin/test -v addons-382533:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 -d /var/lib: (4.766217355s)
I0910 17:30:12.496437 14014 oci.go:107] Successfully prepared a docker volume addons-382533
I0910 17:30:12.496455 14014 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0910 17:30:12.496474 14014 kic.go:194] Starting extracting preloaded images to volume ...
I0910 17:30:12.496540 14014 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19598-5816/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-382533:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 -I lz4 -xf /preloaded.tar -C /extractDir
I0910 17:30:16.385321 14014 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19598-5816/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-382533:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 -I lz4 -xf /preloaded.tar -C /extractDir: (3.888745833s)
I0910 17:30:16.385348 14014 kic.go:203] duration metric: took 3.888871451s to extract preloaded images to volume ...
W0910 17:30:16.385474 14014 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0910 17:30:16.385566 14014 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0910 17:30:16.432925 14014 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-382533 --name addons-382533 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-382533 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-382533 --network addons-382533 --ip 192.168.49.2 --volume addons-382533:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9
I0910 17:30:16.729882 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Running}}
I0910 17:30:16.747151 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:16.766081 14014 cli_runner.go:164] Run: docker exec addons-382533 stat /var/lib/dpkg/alternatives/iptables
I0910 17:30:16.808216 14014 oci.go:144] the created container "addons-382533" has a running status.
I0910 17:30:16.808258 14014 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa...
I0910 17:30:16.891027 14014 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0910 17:30:16.911123 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:16.929677 14014 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0910 17:30:16.929696 14014 kic_runner.go:114] Args: [docker exec --privileged addons-382533 chown docker:docker /home/docker/.ssh/authorized_keys]
I0910 17:30:16.972925 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:16.989777 14014 machine.go:93] provisionDockerMachine start ...
I0910 17:30:16.989865 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:17.009554 14014 main.go:141] libmachine: Using SSH client type: native
I0910 17:30:17.009767 14014 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x8375c0] 0x83a320 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0910 17:30:17.009783 14014 main.go:141] libmachine: About to run SSH command:
hostname
I0910 17:30:17.010405 14014 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: read tcp 127.0.0.1:35836->127.0.0.1:32768: read: connection reset by peer
I0910 17:30:20.134112 14014 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-382533
I0910 17:30:20.134139 14014 ubuntu.go:169] provisioning hostname "addons-382533"
I0910 17:30:20.134199 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:20.149977 14014 main.go:141] libmachine: Using SSH client type: native
I0910 17:30:20.150182 14014 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x8375c0] 0x83a320 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0910 17:30:20.150205 14014 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-382533 && echo "addons-382533" | sudo tee /etc/hostname
I0910 17:30:20.280623 14014 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-382533
I0910 17:30:20.280683 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:20.296593 14014 main.go:141] libmachine: Using SSH client type: native
I0910 17:30:20.296752 14014 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x8375c0] 0x83a320 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0910 17:30:20.296768 14014 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-382533' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-382533/g' /etc/hosts;
else
echo '127.0.1.1 addons-382533' | sudo tee -a /etc/hosts;
fi
fi
I0910 17:30:20.418544 14014 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0910 17:30:20.418583 14014 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19598-5816/.minikube CaCertPath:/home/jenkins/minikube-integration/19598-5816/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19598-5816/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19598-5816/.minikube}
I0910 17:30:20.418629 14014 ubuntu.go:177] setting up certificates
I0910 17:30:20.418651 14014 provision.go:84] configureAuth start
I0910 17:30:20.418716 14014 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-382533
I0910 17:30:20.434221 14014 provision.go:143] copyHostCerts
I0910 17:30:20.434295 14014 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19598-5816/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19598-5816/.minikube/ca.pem (1078 bytes)
I0910 17:30:20.434400 14014 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19598-5816/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19598-5816/.minikube/cert.pem (1123 bytes)
I0910 17:30:20.434459 14014 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19598-5816/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19598-5816/.minikube/key.pem (1679 bytes)
I0910 17:30:20.434512 14014 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19598-5816/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19598-5816/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19598-5816/.minikube/certs/ca-key.pem org=jenkins.addons-382533 san=[127.0.0.1 192.168.49.2 addons-382533 localhost minikube]
I0910 17:30:20.547378 14014 provision.go:177] copyRemoteCerts
I0910 17:30:20.547428 14014 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0910 17:30:20.547461 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:20.563248 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:20.654905 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0910 17:30:20.675530 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0910 17:30:20.696107 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0910 17:30:20.717007 14014 provision.go:87] duration metric: took 298.341794ms to configureAuth
I0910 17:30:20.717031 14014 ubuntu.go:193] setting minikube options for container-runtime
I0910 17:30:20.717181 14014 config.go:182] Loaded profile config "addons-382533": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0910 17:30:20.717220 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:20.733131 14014 main.go:141] libmachine: Using SSH client type: native
I0910 17:30:20.733310 14014 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x8375c0] 0x83a320 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0910 17:30:20.733325 14014 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0910 17:30:20.854866 14014 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0910 17:30:20.854894 14014 ubuntu.go:71] root file system type: overlay
I0910 17:30:20.855043 14014 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0910 17:30:20.855114 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:20.870754 14014 main.go:141] libmachine: Using SSH client type: native
I0910 17:30:20.870936 14014 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x8375c0] 0x83a320 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0910 17:30:20.871004 14014 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0910 17:30:21.000703 14014 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0910 17:30:21.000778 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:21.017943 14014 main.go:141] libmachine: Using SSH client type: native
I0910 17:30:21.018175 14014 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x8375c0] 0x83a320 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0910 17:30:21.018203 14014 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0910 17:30:21.693431 14014 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-09-06 12:06:41.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-10 17:30:20.998790230 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0910 17:30:21.693462 14014 machine.go:96] duration metric: took 4.703666268s to provisionDockerMachine
I0910 17:30:21.693474 14014 client.go:171] duration metric: took 14.356798874s to LocalClient.Create
I0910 17:30:21.693490 14014 start.go:167] duration metric: took 14.356863556s to libmachine.API.Create "addons-382533"
I0910 17:30:21.693497 14014 start.go:293] postStartSetup for "addons-382533" (driver="docker")
I0910 17:30:21.693506 14014 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0910 17:30:21.693555 14014 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0910 17:30:21.693595 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:21.709931 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:21.799022 14014 ssh_runner.go:195] Run: cat /etc/os-release
I0910 17:30:21.801950 14014 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0910 17:30:21.801977 14014 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0910 17:30:21.801985 14014 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0910 17:30:21.801991 14014 info.go:137] Remote host: Ubuntu 22.04.4 LTS
I0910 17:30:21.802001 14014 filesync.go:126] Scanning /home/jenkins/minikube-integration/19598-5816/.minikube/addons for local assets ...
I0910 17:30:21.802057 14014 filesync.go:126] Scanning /home/jenkins/minikube-integration/19598-5816/.minikube/files for local assets ...
I0910 17:30:21.802082 14014 start.go:296] duration metric: took 108.579831ms for postStartSetup
I0910 17:30:21.802350 14014 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-382533
I0910 17:30:21.818402 14014 profile.go:143] Saving config to /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/config.json ...
I0910 17:30:21.818654 14014 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0910 17:30:21.818691 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:21.834763 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:21.919223 14014 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0910 17:30:21.922940 14014 start.go:128] duration metric: took 14.588355416s to createHost
I0910 17:30:21.922962 14014 start.go:83] releasing machines lock for "addons-382533", held for 14.588490086s
I0910 17:30:21.923027 14014 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-382533
I0910 17:30:21.939310 14014 ssh_runner.go:195] Run: cat /version.json
I0910 17:30:21.939353 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:21.939405 14014 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0910 17:30:21.939471 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:21.955693 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:21.956051 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:22.038074 14014 ssh_runner.go:195] Run: systemctl --version
I0910 17:30:22.041796 14014 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0910 17:30:22.111303 14014 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0910 17:30:22.132557 14014 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0910 17:30:22.132613 14014 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0910 17:30:22.156506 14014 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0910 17:30:22.156535 14014 start.go:495] detecting cgroup driver to use...
I0910 17:30:22.156566 14014 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0910 17:30:22.156671 14014 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0910 17:30:22.170581 14014 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0910 17:30:22.179196 14014 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0910 17:30:22.187762 14014 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0910 17:30:22.187811 14014 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0910 17:30:22.196097 14014 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0910 17:30:22.204014 14014 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0910 17:30:22.212002 14014 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0910 17:30:22.220280 14014 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0910 17:30:22.228020 14014 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0910 17:30:22.236761 14014 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0910 17:30:22.245394 14014 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0910 17:30:22.254472 14014 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0910 17:30:22.262143 14014 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0910 17:30:22.269406 14014 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0910 17:30:22.346606 14014 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0910 17:30:22.432613 14014 start.go:495] detecting cgroup driver to use...
I0910 17:30:22.432659 14014 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0910 17:30:22.432706 14014 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0910 17:30:22.443209 14014 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0910 17:30:22.443271 14014 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0910 17:30:22.453944 14014 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0910 17:30:22.468777 14014 ssh_runner.go:195] Run: which cri-dockerd
I0910 17:30:22.471996 14014 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0910 17:30:22.480586 14014 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0910 17:30:22.496333 14014 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0910 17:30:22.581942 14014 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0910 17:30:22.681850 14014 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0910 17:30:22.681982 14014 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0910 17:30:22.698007 14014 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0910 17:30:22.770100 14014 ssh_runner.go:195] Run: sudo systemctl restart docker
I0910 17:30:23.016363 14014 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0910 17:30:23.026817 14014 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0910 17:30:23.036794 14014 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0910 17:30:23.110235 14014 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0910 17:30:23.181491 14014 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0910 17:30:23.255077 14014 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0910 17:30:23.267121 14014 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0910 17:30:23.277280 14014 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0910 17:30:23.359530 14014 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0910 17:30:23.417137 14014 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0910 17:30:23.417206 14014 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0910 17:30:23.420928 14014 start.go:563] Will wait 60s for crictl version
I0910 17:30:23.420979 14014 ssh_runner.go:195] Run: which crictl
I0910 17:30:23.423987 14014 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0910 17:30:23.454929 14014 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.2.1
RuntimeApiVersion: v1
I0910 17:30:23.454981 14014 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0910 17:30:23.477131 14014 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0910 17:30:23.500519 14014 out.go:235] * Preparing Kubernetes v1.31.0 on Docker 27.2.1 ...
I0910 17:30:23.500585 14014 cli_runner.go:164] Run: docker network inspect addons-382533 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0910 17:30:23.514945 14014 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0910 17:30:23.518239 14014 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0910 17:30:23.527726 14014 kubeadm.go:883] updating cluster {Name:addons-382533 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-382533 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0910 17:30:23.527820 14014 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0910 17:30:23.527860 14014 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0910 17:30:23.545062 14014 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-controller-manager:v1.31.0
registry.k8s.io/kube-scheduler:v1.31.0
registry.k8s.io/kube-apiserver:v1.31.0
registry.k8s.io/kube-proxy:v1.31.0
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
registry.k8s.io/coredns/coredns:v1.11.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0910 17:30:23.545083 14014 docker.go:615] Images already preloaded, skipping extraction
I0910 17:30:23.545157 14014 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0910 17:30:23.562345 14014 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-controller-manager:v1.31.0
registry.k8s.io/kube-scheduler:v1.31.0
registry.k8s.io/kube-apiserver:v1.31.0
registry.k8s.io/kube-proxy:v1.31.0
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
registry.k8s.io/coredns/coredns:v1.11.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0910 17:30:23.562373 14014 cache_images.go:84] Images are preloaded, skipping loading
I0910 17:30:23.562390 14014 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.0 docker true true} ...
I0910 17:30:23.562475 14014 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.0/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-382533 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.0 ClusterName:addons-382533 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0910 17:30:23.562523 14014 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0910 17:30:23.603210 14014 cni.go:84] Creating CNI manager for ""
I0910 17:30:23.603243 14014 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0910 17:30:23.603259 14014 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0910 17:30:23.603283 14014 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.0 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-382533 NodeName:addons-382533 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0910 17:30:23.603435 14014 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-382533"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.0
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0910 17:30:23.603499 14014 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.0
I0910 17:30:23.611172 14014 binaries.go:44] Found k8s binaries, skipping transfer
I0910 17:30:23.611228 14014 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0910 17:30:23.618366 14014 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0910 17:30:23.633328 14014 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0910 17:30:23.648330 14014 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0910 17:30:23.663470 14014 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0910 17:30:23.666478 14014 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0910 17:30:23.675977 14014 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0910 17:30:23.751701 14014 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0910 17:30:23.763994 14014 certs.go:68] Setting up /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533 for IP: 192.168.49.2
I0910 17:30:23.764025 14014 certs.go:194] generating shared ca certs ...
I0910 17:30:23.764047 14014 certs.go:226] acquiring lock for ca certs: {Name:mk81ad856ddc9549fa70e05ee706a547d400822f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:23.764170 14014 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19598-5816/.minikube/ca.key
I0910 17:30:23.891692 14014 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19598-5816/.minikube/ca.crt ...
I0910 17:30:23.891717 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/ca.crt: {Name:mkd98554b29d56bc875ce122a07e6b39e771aff9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:23.891865 14014 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19598-5816/.minikube/ca.key ...
I0910 17:30:23.891875 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/ca.key: {Name:mk7eadea766f8522786887092bab7223b5a5936b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:23.891950 14014 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19598-5816/.minikube/proxy-client-ca.key
I0910 17:30:24.116495 14014 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19598-5816/.minikube/proxy-client-ca.crt ...
I0910 17:30:24.116525 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/proxy-client-ca.crt: {Name:mkace679209945cf5e9edfa0001a48433d82bd72 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:24.116685 14014 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19598-5816/.minikube/proxy-client-ca.key ...
I0910 17:30:24.116695 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/proxy-client-ca.key: {Name:mk2eaef06c3b5d248fcd4643ee84c0f506afe9d4 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:24.116761 14014 certs.go:256] generating profile certs ...
I0910 17:30:24.116812 14014 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/client.key
I0910 17:30:24.116833 14014 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/client.crt with IP's: []
I0910 17:30:24.392213 14014 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/client.crt ...
I0910 17:30:24.392239 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/client.crt: {Name:mke62895fe2e3c1b832bf46dedd0870b6c07174d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:24.392395 14014 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/client.key ...
I0910 17:30:24.392409 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/client.key: {Name:mk9e78b54c84edd53efe1972dc01a89162b044cf Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:24.392470 14014 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.key.fa87e0ac
I0910 17:30:24.392486 14014 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.crt.fa87e0ac with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0910 17:30:24.491497 14014 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.crt.fa87e0ac ...
I0910 17:30:24.491522 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.crt.fa87e0ac: {Name:mk448133abac09c65cbe4bb16dc35ede911c974d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:24.491681 14014 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.key.fa87e0ac ...
I0910 17:30:24.491694 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.key.fa87e0ac: {Name:mk7afda8772350c638816c66b441c8245a7ff9ca Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:24.491759 14014 certs.go:381] copying /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.crt.fa87e0ac -> /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.crt
I0910 17:30:24.491828 14014 certs.go:385] copying /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.key.fa87e0ac -> /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.key
I0910 17:30:24.491872 14014 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/proxy-client.key
I0910 17:30:24.491888 14014 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/proxy-client.crt with IP's: []
I0910 17:30:24.590604 14014 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/proxy-client.crt ...
I0910 17:30:24.590630 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/proxy-client.crt: {Name:mk8cf21af8478c928b8779649c7e2179a068abcf Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:24.590791 14014 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/proxy-client.key ...
I0910 17:30:24.590801 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/proxy-client.key: {Name:mk547a71a82ee11eaba6aea239968cef9c6b7b31 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:24.591002 14014 certs.go:484] found cert: /home/jenkins/minikube-integration/19598-5816/.minikube/certs/ca-key.pem (1675 bytes)
I0910 17:30:24.591034 14014 certs.go:484] found cert: /home/jenkins/minikube-integration/19598-5816/.minikube/certs/ca.pem (1078 bytes)
I0910 17:30:24.591056 14014 certs.go:484] found cert: /home/jenkins/minikube-integration/19598-5816/.minikube/certs/cert.pem (1123 bytes)
I0910 17:30:24.591077 14014 certs.go:484] found cert: /home/jenkins/minikube-integration/19598-5816/.minikube/certs/key.pem (1679 bytes)
I0910 17:30:24.591616 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0910 17:30:24.613288 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0910 17:30:24.634063 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0910 17:30:24.654284 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I0910 17:30:24.675155 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0910 17:30:24.695492 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0910 17:30:24.715848 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0910 17:30:24.735810 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/profiles/addons-382533/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0910 17:30:24.755776 14014 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19598-5816/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0910 17:30:24.775911 14014 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0910 17:30:24.790900 14014 ssh_runner.go:195] Run: openssl version
I0910 17:30:24.795628 14014 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0910 17:30:24.803413 14014 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0910 17:30:24.806217 14014 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 10 17:30 /usr/share/ca-certificates/minikubeCA.pem
I0910 17:30:24.806276 14014 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0910 17:30:24.812111 14014 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0910 17:30:24.820065 14014 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0910 17:30:24.822925 14014 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0910 17:30:24.822965 14014 kubeadm.go:392] StartCluster: {Name:addons-382533 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1725963390-19606@sha256:05c3fb4a3ac73e1a547cb186e5aec949a4a3d18af7d1444e0d1365c17dbedef9 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-382533 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0910 17:30:24.823051 14014 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0910 17:30:24.838782 14014 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0910 17:30:24.846222 14014 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0910 17:30:24.853670 14014 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0910 17:30:24.853714 14014 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0910 17:30:24.861040 14014 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0910 17:30:24.861059 14014 kubeadm.go:157] found existing configuration files:
I0910 17:30:24.861100 14014 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0910 17:30:24.868319 14014 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0910 17:30:24.868374 14014 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0910 17:30:24.875108 14014 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0910 17:30:24.881895 14014 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0910 17:30:24.881939 14014 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0910 17:30:24.888436 14014 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0910 17:30:24.895280 14014 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0910 17:30:24.895322 14014 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0910 17:30:24.901853 14014 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0910 17:30:24.908527 14014 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0910 17:30:24.908568 14014 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0910 17:30:24.915249 14014 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.0:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0910 17:30:24.946264 14014 kubeadm.go:310] W0910 17:30:24.945635 1927 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0910 17:30:24.946700 14014 kubeadm.go:310] W0910 17:30:24.946240 1927 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0910 17:30:24.967524 14014 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1067-gcp\n", err: exit status 1
I0910 17:30:25.013322 14014 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0910 17:30:34.023025 14014 kubeadm.go:310] [init] Using Kubernetes version: v1.31.0
I0910 17:30:34.023073 14014 kubeadm.go:310] [preflight] Running pre-flight checks
I0910 17:30:34.023177 14014 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0910 17:30:34.023241 14014 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1067-gcp[0m
I0910 17:30:34.023272 14014 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0910 17:30:34.023314 14014 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0910 17:30:34.023354 14014 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0910 17:30:34.023401 14014 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0910 17:30:34.023441 14014 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0910 17:30:34.023486 14014 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0910 17:30:34.023530 14014 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0910 17:30:34.023568 14014 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0910 17:30:34.023610 14014 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0910 17:30:34.023649 14014 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0910 17:30:34.023712 14014 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0910 17:30:34.023831 14014 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0910 17:30:34.023954 14014 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0910 17:30:34.024038 14014 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0910 17:30:34.025876 14014 out.go:235] - Generating certificates and keys ...
I0910 17:30:34.025959 14014 kubeadm.go:310] [certs] Using existing ca certificate authority
I0910 17:30:34.026032 14014 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0910 17:30:34.026111 14014 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0910 17:30:34.026182 14014 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0910 17:30:34.026273 14014 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0910 17:30:34.026349 14014 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0910 17:30:34.026437 14014 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0910 17:30:34.026567 14014 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-382533 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0910 17:30:34.026619 14014 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0910 17:30:34.026713 14014 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-382533 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0910 17:30:34.026767 14014 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0910 17:30:34.026869 14014 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0910 17:30:34.026926 14014 kubeadm.go:310] [certs] Generating "sa" key and public key
I0910 17:30:34.026974 14014 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0910 17:30:34.027024 14014 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0910 17:30:34.027086 14014 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0910 17:30:34.027160 14014 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0910 17:30:34.027218 14014 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0910 17:30:34.027269 14014 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0910 17:30:34.027343 14014 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0910 17:30:34.027402 14014 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0910 17:30:34.028925 14014 out.go:235] - Booting up control plane ...
I0910 17:30:34.029000 14014 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0910 17:30:34.029062 14014 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0910 17:30:34.029125 14014 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0910 17:30:34.029216 14014 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0910 17:30:34.029293 14014 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0910 17:30:34.029347 14014 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0910 17:30:34.029465 14014 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0910 17:30:34.029560 14014 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0910 17:30:34.029613 14014 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 501.361822ms
I0910 17:30:34.029671 14014 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0910 17:30:34.029725 14014 kubeadm.go:310] [api-check] The API server is healthy after 4.501172034s
I0910 17:30:34.029828 14014 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0910 17:30:34.029943 14014 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0910 17:30:34.029993 14014 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0910 17:30:34.030149 14014 kubeadm.go:310] [mark-control-plane] Marking the node addons-382533 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0910 17:30:34.030201 14014 kubeadm.go:310] [bootstrap-token] Using token: ijg2u1.ou5hsbkerjrrr66c
I0910 17:30:34.031639 14014 out.go:235] - Configuring RBAC rules ...
I0910 17:30:34.031738 14014 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0910 17:30:34.031811 14014 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0910 17:30:34.031926 14014 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0910 17:30:34.032044 14014 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0910 17:30:34.032172 14014 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0910 17:30:34.032299 14014 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0910 17:30:34.032433 14014 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0910 17:30:34.032471 14014 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0910 17:30:34.032513 14014 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0910 17:30:34.032519 14014 kubeadm.go:310]
I0910 17:30:34.032577 14014 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0910 17:30:34.032586 14014 kubeadm.go:310]
I0910 17:30:34.032651 14014 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0910 17:30:34.032662 14014 kubeadm.go:310]
I0910 17:30:34.032686 14014 kubeadm.go:310] mkdir -p $HOME/.kube
I0910 17:30:34.032733 14014 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0910 17:30:34.032775 14014 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0910 17:30:34.032781 14014 kubeadm.go:310]
I0910 17:30:34.032824 14014 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0910 17:30:34.032830 14014 kubeadm.go:310]
I0910 17:30:34.032877 14014 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0910 17:30:34.032884 14014 kubeadm.go:310]
I0910 17:30:34.032926 14014 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0910 17:30:34.032988 14014 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0910 17:30:34.033044 14014 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0910 17:30:34.033050 14014 kubeadm.go:310]
I0910 17:30:34.033117 14014 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0910 17:30:34.033208 14014 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0910 17:30:34.033219 14014 kubeadm.go:310]
I0910 17:30:34.033332 14014 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token ijg2u1.ou5hsbkerjrrr66c \
I0910 17:30:34.033438 14014 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:f66b60f58b622f8c999fe2b8f76480d5d0fa35c469684373fff9166b550dcb65 \
I0910 17:30:34.033470 14014 kubeadm.go:310] --control-plane
I0910 17:30:34.033483 14014 kubeadm.go:310]
I0910 17:30:34.033604 14014 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0910 17:30:34.033611 14014 kubeadm.go:310]
I0910 17:30:34.033714 14014 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token ijg2u1.ou5hsbkerjrrr66c \
I0910 17:30:34.033865 14014 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:f66b60f58b622f8c999fe2b8f76480d5d0fa35c469684373fff9166b550dcb65
I0910 17:30:34.033877 14014 cni.go:84] Creating CNI manager for ""
I0910 17:30:34.033898 14014 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0910 17:30:34.035337 14014 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0910 17:30:34.036516 14014 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0910 17:30:34.044550 14014 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0910 17:30:34.060532 14014 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0910 17:30:34.060608 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:34.060620 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-382533 minikube.k8s.io/updated_at=2024_09_10T17_30_34_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=37b4bace07cd53444288cad630e4db4b688b8c18 minikube.k8s.io/name=addons-382533 minikube.k8s.io/primary=true
I0910 17:30:34.068319 14014 ops.go:34] apiserver oom_adj: -16
I0910 17:30:34.145298 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:34.645420 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:35.145702 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:35.646352 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:36.145758 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:36.645712 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:37.145457 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:37.645914 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:38.145511 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:38.645551 14014 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0910 17:30:38.710665 14014 kubeadm.go:1113] duration metric: took 4.650116215s to wait for elevateKubeSystemPrivileges
I0910 17:30:38.710692 14014 kubeadm.go:394] duration metric: took 13.887730532s to StartCluster
I0910 17:30:38.710708 14014 settings.go:142] acquiring lock: {Name:mk1f43a03dcdaafb06737758e1f83fe909c27ae9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:38.710846 14014 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19598-5816/kubeconfig
I0910 17:30:38.711232 14014 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19598-5816/kubeconfig: {Name:mk1480fbd449361d4db9efc8cf3426d8ceb05753 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0910 17:30:38.711437 14014 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0910 17:30:38.711447 14014 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0910 17:30:38.711547 14014 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0910 17:30:38.711636 14014 config.go:182] Loaded profile config "addons-382533": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0910 17:30:38.711645 14014 addons.go:69] Setting yakd=true in profile "addons-382533"
I0910 17:30:38.711651 14014 addons.go:69] Setting helm-tiller=true in profile "addons-382533"
I0910 17:30:38.711686 14014 addons.go:234] Setting addon yakd=true in "addons-382533"
I0910 17:30:38.711692 14014 addons.go:69] Setting inspektor-gadget=true in profile "addons-382533"
I0910 17:30:38.711696 14014 addons.go:69] Setting ingress=true in profile "addons-382533"
I0910 17:30:38.711716 14014 addons.go:234] Setting addon helm-tiller=true in "addons-382533"
I0910 17:30:38.711724 14014 addons.go:234] Setting addon inspektor-gadget=true in "addons-382533"
I0910 17:30:38.711729 14014 addons.go:234] Setting addon ingress=true in "addons-382533"
I0910 17:30:38.711736 14014 addons.go:69] Setting registry=true in profile "addons-382533"
I0910 17:30:38.711753 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.711690 14014 addons.go:69] Setting ingress-dns=true in profile "addons-382533"
I0910 17:30:38.711762 14014 addons.go:234] Setting addon registry=true in "addons-382533"
I0910 17:30:38.711766 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.711778 14014 addons.go:234] Setting addon ingress-dns=true in "addons-382533"
I0910 17:30:38.711786 14014 addons.go:69] Setting storage-provisioner=true in profile "addons-382533"
I0910 17:30:38.711803 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.711809 14014 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-382533"
I0910 17:30:38.711828 14014 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-382533"
I0910 17:30:38.711943 14014 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-382533"
I0910 17:30:38.712001 14014 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-382533"
I0910 17:30:38.712031 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.712111 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.712259 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.712282 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.712294 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.712319 14014 addons.go:69] Setting cloud-spanner=true in profile "addons-382533"
I0910 17:30:38.712381 14014 addons.go:234] Setting addon cloud-spanner=true in "addons-382533"
I0910 17:30:38.712412 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.712498 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.712659 14014 addons.go:69] Setting volcano=true in profile "addons-382533"
I0910 17:30:38.712701 14014 addons.go:234] Setting addon volcano=true in "addons-382533"
I0910 17:30:38.712715 14014 addons.go:69] Setting default-storageclass=true in profile "addons-382533"
I0910 17:30:38.712730 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.712739 14014 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-382533"
I0910 17:30:38.712835 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.712986 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.713162 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.715716 14014 addons.go:69] Setting metrics-server=true in profile "addons-382533"
I0910 17:30:38.715758 14014 addons.go:234] Setting addon metrics-server=true in "addons-382533"
I0910 17:30:38.715793 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.716272 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.711729 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.716845 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.717372 14014 out.go:177] * Verifying Kubernetes components...
I0910 17:30:38.718035 14014 addons.go:69] Setting volumesnapshots=true in profile "addons-382533"
I0910 17:30:38.718068 14014 addons.go:234] Setting addon volumesnapshots=true in "addons-382533"
I0910 17:30:38.717602 14014 addons.go:69] Setting gcp-auth=true in profile "addons-382533"
I0910 17:30:38.711759 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.718140 14014 mustload.go:65] Loading cluster: addons-382533
I0910 17:30:38.711804 14014 addons.go:234] Setting addon storage-provisioner=true in "addons-382533"
I0910 17:30:38.711782 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.718360 14014 config.go:182] Loaded profile config "addons-382533": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0910 17:30:38.718589 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.718606 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.718627 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.718698 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.718096 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.711674 14014 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-382533"
I0910 17:30:38.719159 14014 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-382533"
I0910 17:30:38.719191 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.719627 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.723443 14014 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0910 17:30:38.739232 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.739402 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.750494 14014 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-382533"
I0910 17:30:38.750543 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.751046 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.772240 14014 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0910 17:30:38.774277 14014 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0910 17:30:38.775910 14014 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0910 17:30:38.776008 14014 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0910 17:30:38.777256 14014 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0910 17:30:38.777274 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0910 17:30:38.777328 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.778130 14014 addons.go:234] Setting addon default-storageclass=true in "addons-382533"
I0910 17:30:38.778183 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.778663 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:38.779158 14014 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0910 17:30:38.779261 14014 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0910 17:30:38.780566 14014 out.go:177] - Using image docker.io/registry:2.8.3
I0910 17:30:38.780678 14014 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0910 17:30:38.781854 14014 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0910 17:30:38.781934 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0910 17:30:38.781973 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.782166 14014 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0910 17:30:38.782394 14014 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0910 17:30:38.783547 14014 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0910 17:30:38.783563 14014 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0910 17:30:38.783609 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.783780 14014 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0910 17:30:38.785287 14014 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0910 17:30:38.786921 14014 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0910 17:30:38.786939 14014 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0910 17:30:38.786985 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.793690 14014 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0910 17:30:38.796740 14014 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0910 17:30:38.796831 14014 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0910 17:30:38.802736 14014 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0910 17:30:38.802756 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0910 17:30:38.802823 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.802977 14014 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0910 17:30:38.805218 14014 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0910 17:30:38.805222 14014 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0910 17:30:38.805408 14014 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0910 17:30:38.805419 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0910 17:30:38.805478 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.813542 14014 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0910 17:30:38.813573 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0910 17:30:38.813636 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.814940 14014 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0910 17:30:38.814964 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0910 17:30:38.815019 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.820034 14014 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0910 17:30:38.823903 14014 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0910 17:30:38.824078 14014 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0910 17:30:38.824115 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0910 17:30:38.824177 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.828285 14014 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0910 17:30:38.830165 14014 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0910 17:30:38.830189 14014 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0910 17:30:38.830247 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.830257 14014 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0910 17:30:38.830408 14014 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0910 17:30:38.831944 14014 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0910 17:30:38.831969 14014 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0910 17:30:38.832026 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.834060 14014 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0910 17:30:38.836309 14014 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0910 17:30:38.836331 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0910 17:30:38.836393 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.841485 14014 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0910 17:30:38.847008 14014 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0910 17:30:38.848518 14014 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0910 17:30:38.848539 14014 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0910 17:30:38.848615 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.848830 14014 out.go:177] - Using image docker.io/busybox:stable
I0910 17:30:38.850165 14014 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0910 17:30:38.850179 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0910 17:30:38.850224 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.856616 14014 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0910 17:30:38.856642 14014 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0910 17:30:38.856723 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:38.857256 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.860897 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.862029 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.867038 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.881254 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:38.885035 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.888335 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.899059 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.899920 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.904234 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.904608 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.905961 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.906911 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.908118 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.912013 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:38.920343 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
W0910 17:30:38.947215 14014 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0910 17:30:38.947249 14014 retry.go:31] will retry after 238.544227ms: ssh: handshake failed: EOF
I0910 17:30:39.063698 14014 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0910 17:30:39.063745 14014 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0910 17:30:39.256027 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0910 17:30:39.356662 14014 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0910 17:30:39.356694 14014 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0910 17:30:39.454028 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0910 17:30:39.455972 14014 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0910 17:30:39.456002 14014 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0910 17:30:39.456380 14014 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0910 17:30:39.456400 14014 ssh_runner.go:362] scp helm-tiller/helm-tiller-rbac.yaml --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0910 17:30:39.457707 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0910 17:30:39.464839 14014 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0910 17:30:39.464865 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0910 17:30:39.544613 14014 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0910 17:30:39.544642 14014 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0910 17:30:39.563099 14014 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0910 17:30:39.563133 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0910 17:30:39.643768 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0910 17:30:39.644306 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0910 17:30:39.655484 14014 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0910 17:30:39.655510 14014 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0910 17:30:39.657542 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0910 17:30:39.661305 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0910 17:30:39.669063 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0910 17:30:39.751411 14014 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0910 17:30:39.751487 14014 ssh_runner.go:362] scp helm-tiller/helm-tiller-svc.yaml --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0910 17:30:39.753051 14014 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0910 17:30:39.753111 14014 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0910 17:30:39.756548 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0910 17:30:39.764926 14014 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0910 17:30:39.764949 14014 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0910 17:30:39.865407 14014 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0910 17:30:39.865483 14014 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0910 17:30:39.950203 14014 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0910 17:30:39.950230 14014 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0910 17:30:39.957914 14014 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0910 17:30:39.957940 14014 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0910 17:30:40.152446 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0910 17:30:40.155351 14014 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0910 17:30:40.155384 14014 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0910 17:30:40.256292 14014 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0910 17:30:40.256322 14014 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0910 17:30:40.353937 14014 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0910 17:30:40.353965 14014 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0910 17:30:40.546199 14014 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0910 17:30:40.546227 14014 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0910 17:30:40.549560 14014 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0910 17:30:40.549582 14014 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0910 17:30:40.552126 14014 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0910 17:30:40.552146 14014 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0910 17:30:40.566990 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0910 17:30:40.953771 14014 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0910 17:30:40.953802 14014 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0910 17:30:41.050112 14014 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.986253424s)
I0910 17:30:41.051180 14014 node_ready.go:35] waiting up to 6m0s for node "addons-382533" to be "Ready" ...
I0910 17:30:41.051404 14014 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.987616658s)
I0910 17:30:41.051433 14014 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0910 17:30:41.054149 14014 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0910 17:30:41.054202 14014 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0910 17:30:41.054693 14014 node_ready.go:49] node "addons-382533" has status "Ready":"True"
I0910 17:30:41.054716 14014 node_ready.go:38] duration metric: took 3.506097ms for node "addons-382533" to be "Ready" ...
I0910 17:30:41.054728 14014 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0910 17:30:41.065611 14014 pod_ready.go:79] waiting up to 6m0s for pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace to be "Ready" ...
I0910 17:30:41.249845 14014 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0910 17:30:41.249872 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0910 17:30:41.357191 14014 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0910 17:30:41.357218 14014 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0910 17:30:41.461617 14014 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0910 17:30:41.461640 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0910 17:30:41.555286 14014 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-382533" context rescaled to 1 replicas
I0910 17:30:41.751873 14014 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0910 17:30:41.751903 14014 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0910 17:30:42.050022 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0910 17:30:42.161283 14014 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0910 17:30:42.161386 14014 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0910 17:30:42.356163 14014 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0910 17:30:42.356243 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0910 17:30:42.443328 14014 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0910 17:30:42.443356 14014 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0910 17:30:42.843927 14014 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0910 17:30:42.843963 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0910 17:30:42.856069 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (3.599998071s)
I0910 17:30:42.856164 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (3.402110399s)
I0910 17:30:42.856212 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (3.398477684s)
I0910 17:30:42.943701 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0910 17:30:42.945203 14014 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0910 17:30:42.945235 14014 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0910 17:30:43.152106 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:30:43.160618 14014 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0910 17:30:43.160703 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0910 17:30:43.462240 14014 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0910 17:30:43.462351 14014 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0910 17:30:43.565176 14014 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0910 17:30:43.565256 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0910 17:30:44.067228 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0910 17:30:44.155610 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0910 17:30:44.248849 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (4.605039396s)
I0910 17:30:44.249117 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (4.604782738s)
I0910 17:30:44.249359 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (4.591744487s)
I0910 17:30:44.249395 14014 addons.go:475] Verifying addon registry=true in "addons-382533"
I0910 17:30:44.249686 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (4.588312797s)
I0910 17:30:44.251148 14014 out.go:177] * Verifying registry addon...
I0910 17:30:44.253633 14014 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
W0910 17:30:44.258934 14014 out.go:270] ! Enabling 'storage-provisioner-rancher' returned an error: running callbacks: [Error making local-path the default storage class: Error while marking storage class local-path as default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
I0910 17:30:44.260623 14014 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0910 17:30:44.260681 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:44.764395 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:45.155628 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:30:45.257955 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:45.758555 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:45.950547 14014 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0910 17:30:45.950628 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:45.981230 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:46.259297 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:46.758311 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:47.046847 14014 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0910 17:30:47.258936 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:47.343553 14014 addons.go:234] Setting addon gcp-auth=true in "addons-382533"
I0910 17:30:47.343610 14014 host.go:66] Checking if "addons-382533" exists ...
I0910 17:30:47.344122 14014 cli_runner.go:164] Run: docker container inspect addons-382533 --format={{.State.Status}}
I0910 17:30:47.374376 14014 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0910 17:30:47.374423 14014 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-382533
I0910 17:30:47.395881 14014 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19598-5816/.minikube/machines/addons-382533/id_rsa Username:docker}
I0910 17:30:47.647317 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:30:47.757735 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:48.259444 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:48.846477 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:49.261383 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:49.747991 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:30:49.761313 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:50.350956 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:50.758146 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:51.066437 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.397338956s)
I0910 17:30:51.066563 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (11.309949181s)
I0910 17:30:51.066592 14014 addons.go:475] Verifying addon ingress=true in "addons-382533"
I0910 17:30:51.066680 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (10.914201652s)
I0910 17:30:51.066765 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (10.499745276s)
I0910 17:30:51.066783 14014 addons.go:475] Verifying addon metrics-server=true in "addons-382533"
I0910 17:30:51.066851 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (9.01675229s)
I0910 17:30:51.067057 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (8.123315562s)
W0910 17:30:51.067086 14014 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0910 17:30:51.067106 14014 retry.go:31] will retry after 172.249594ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0910 17:30:51.068207 14014 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-382533 service yakd-dashboard -n yakd-dashboard
I0910 17:30:51.068382 14014 out.go:177] * Verifying ingress addon...
I0910 17:30:51.145607 14014 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0910 17:30:51.153483 14014 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0910 17:30:51.153513 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:51.240549 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0910 17:30:51.260058 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:51.650838 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:51.762897 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:52.071041 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:30:52.150333 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:52.258116 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:52.650610 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:52.655233 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (8.587903395s)
I0910 17:30:52.655267 14014 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-382533"
I0910 17:30:52.655534 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (8.499697487s)
I0910 17:30:52.655613 14014 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.281213211s)
I0910 17:30:52.656812 14014 out.go:177] * Verifying csi-hostpath-driver addon...
I0910 17:30:52.656812 14014 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0910 17:30:52.658679 14014 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0910 17:30:52.660044 14014 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0910 17:30:52.660533 14014 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0910 17:30:52.660561 14014 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0910 17:30:52.672998 14014 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0910 17:30:52.673030 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:52.757950 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:52.758158 14014 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0910 17:30:52.758175 14014 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0910 17:30:52.781201 14014 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0910 17:30:52.781224 14014 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0910 17:30:52.865836 14014 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0910 17:30:53.151095 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:53.165570 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:53.258044 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:53.650254 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:53.665156 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:53.748739 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.508138313s)
I0910 17:30:53.757565 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:54.151342 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:54.172971 14014 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.307088959s)
I0910 17:30:54.174593 14014 addons.go:475] Verifying addon gcp-auth=true in "addons-382533"
I0910 17:30:54.176108 14014 out.go:177] * Verifying gcp-auth addon...
I0910 17:30:54.178000 14014 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0910 17:30:54.250715 14014 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0910 17:30:54.252620 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:54.257269 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:54.571098 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:30:54.649572 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:54.664578 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:54.757433 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:55.150330 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:55.250672 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:55.256704 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:55.650633 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:55.664675 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:55.756824 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:56.150442 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:56.164347 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:56.257731 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:56.571656 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:30:56.650279 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:56.664396 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:56.757519 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:57.149781 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:57.164752 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:57.256844 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:57.650024 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:57.664340 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:57.757499 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:58.149442 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:58.164125 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:58.257554 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:58.572117 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:30:58.650024 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:58.664197 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:58.756752 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:59.149909 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:59.164643 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:59.257195 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:30:59.649035 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:30:59.663566 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:30:59.757613 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:00.149439 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:00.164012 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:00.256896 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:00.650144 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:00.665902 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:00.757188 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:01.071141 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:31:01.150067 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:01.164161 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:01.257595 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:01.649827 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:01.664141 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:01.757679 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:02.149311 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:02.164490 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:02.256908 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:02.650340 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:02.664623 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:02.756919 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:03.071920 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:31:03.150780 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:03.164688 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:03.257374 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:03.649822 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:03.665086 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:03.757796 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:04.150591 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:04.164487 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:04.257753 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:04.650125 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:04.664163 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:04.802270 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:05.157168 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:05.159349 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:31:05.164497 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:05.257174 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:05.650445 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:05.664030 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:05.757239 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:06.149822 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:06.163707 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:06.256654 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:06.649748 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:06.664370 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:06.757902 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:07.149663 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:07.164128 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:07.256715 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:07.571247 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:31:07.649889 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:07.664720 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:07.757054 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:08.150054 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:08.164510 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:08.256654 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:08.649360 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:08.664057 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:08.757188 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:09.149781 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:09.164445 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:09.257547 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:09.649330 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:09.663708 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:09.756568 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:10.071762 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:31:10.150407 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:10.164356 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:10.257753 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:10.650041 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:10.664129 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:10.757377 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:11.240821 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:11.241097 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:11.340706 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:11.649711 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:11.664507 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:11.757311 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:12.149672 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:12.164425 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:12.257310 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:12.570745 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:31:12.650448 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:12.664187 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:12.758285 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:13.150222 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:13.163400 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:13.257463 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:13.649843 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:13.663675 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:13.756684 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:14.149511 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:14.164160 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:14.258032 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:14.571396 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:31:14.650290 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:14.664769 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:14.757417 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:15.151071 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:15.165441 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:15.258313 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:15.650106 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:15.664235 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:15.757406 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:16.150046 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:16.165045 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:16.257366 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:16.649758 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:16.664997 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:16.757178 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:17.071186 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:31:17.149916 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:17.164403 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:17.257474 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:17.649901 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:17.663491 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:17.756860 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0910 17:31:18.149781 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:18.164551 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:18.257081 14014 kapi.go:107] duration metric: took 34.003449669s to wait for kubernetes.io/minikube-addons=registry ...
I0910 17:31:18.649792 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:18.665101 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:19.150031 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:19.163801 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:19.571040 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:31:19.648980 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:19.663684 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:20.150030 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:20.163721 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:20.652355 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:20.663613 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:21.149558 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:21.164510 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:21.571398 14014 pod_ready.go:103] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"False"
I0910 17:31:21.650320 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:21.664422 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:22.071316 14014 pod_ready.go:93] pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace has status "Ready":"True"
I0910 17:31:22.071339 14014 pod_ready.go:82] duration metric: took 41.00564864s for pod "coredns-6f6b679f8f-g2x6g" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.071349 14014 pod_ready.go:79] waiting up to 6m0s for pod "coredns-6f6b679f8f-gft7v" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.072778 14014 pod_ready.go:98] error getting pod "coredns-6f6b679f8f-gft7v" in "kube-system" namespace (skipping!): pods "coredns-6f6b679f8f-gft7v" not found
I0910 17:31:22.072801 14014 pod_ready.go:82] duration metric: took 1.445536ms for pod "coredns-6f6b679f8f-gft7v" in "kube-system" namespace to be "Ready" ...
E0910 17:31:22.072814 14014 pod_ready.go:67] WaitExtra: waitPodCondition: error getting pod "coredns-6f6b679f8f-gft7v" in "kube-system" namespace (skipping!): pods "coredns-6f6b679f8f-gft7v" not found
I0910 17:31:22.072823 14014 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-382533" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.076994 14014 pod_ready.go:93] pod "etcd-addons-382533" in "kube-system" namespace has status "Ready":"True"
I0910 17:31:22.077016 14014 pod_ready.go:82] duration metric: took 4.182798ms for pod "etcd-addons-382533" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.077025 14014 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-382533" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.080820 14014 pod_ready.go:93] pod "kube-apiserver-addons-382533" in "kube-system" namespace has status "Ready":"True"
I0910 17:31:22.080837 14014 pod_ready.go:82] duration metric: took 3.806365ms for pod "kube-apiserver-addons-382533" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.080846 14014 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-382533" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.084688 14014 pod_ready.go:93] pod "kube-controller-manager-addons-382533" in "kube-system" namespace has status "Ready":"True"
I0910 17:31:22.084711 14014 pod_ready.go:82] duration metric: took 3.857769ms for pod "kube-controller-manager-addons-382533" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.084724 14014 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-29nvr" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.149955 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:22.163533 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:22.269214 14014 pod_ready.go:93] pod "kube-proxy-29nvr" in "kube-system" namespace has status "Ready":"True"
I0910 17:31:22.269235 14014 pod_ready.go:82] duration metric: took 184.502604ms for pod "kube-proxy-29nvr" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.269246 14014 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-382533" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.649712 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:22.664301 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:22.668333 14014 pod_ready.go:93] pod "kube-scheduler-addons-382533" in "kube-system" namespace has status "Ready":"True"
I0910 17:31:22.668352 14014 pod_ready.go:82] duration metric: took 399.099198ms for pod "kube-scheduler-addons-382533" in "kube-system" namespace to be "Ready" ...
I0910 17:31:22.668362 14014 pod_ready.go:39] duration metric: took 41.613616955s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0910 17:31:22.668383 14014 api_server.go:52] waiting for apiserver process to appear ...
I0910 17:31:22.668441 14014 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0910 17:31:22.682139 14014 api_server.go:72] duration metric: took 43.97066723s to wait for apiserver process to appear ...
I0910 17:31:22.682160 14014 api_server.go:88] waiting for apiserver healthz status ...
I0910 17:31:22.682184 14014 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0910 17:31:22.685470 14014 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0910 17:31:22.686185 14014 api_server.go:141] control plane version: v1.31.0
I0910 17:31:22.686205 14014 api_server.go:131] duration metric: took 4.038861ms to wait for apiserver health ...
I0910 17:31:22.686214 14014 system_pods.go:43] waiting for kube-system pods to appear ...
I0910 17:31:22.873606 14014 system_pods.go:59] 18 kube-system pods found
I0910 17:31:22.873636 14014 system_pods.go:61] "coredns-6f6b679f8f-g2x6g" [9ad2f9ec-61cf-4a29-a8ff-24e2bd3088db] Running
I0910 17:31:22.873645 14014 system_pods.go:61] "csi-hostpath-attacher-0" [06d0a0cc-d0f6-4e05-8af6-bf4f4d2d20fc] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0910 17:31:22.873652 14014 system_pods.go:61] "csi-hostpath-resizer-0" [dd5662db-6ebf-48a7-ad31-c70fad0c33fb] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0910 17:31:22.873660 14014 system_pods.go:61] "csi-hostpathplugin-9jbqt" [1b107483-67fb-46fb-8d25-9aa28471a833] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0910 17:31:22.873665 14014 system_pods.go:61] "etcd-addons-382533" [c444f86b-711d-4fc5-89eb-20cea8837776] Running
I0910 17:31:22.873669 14014 system_pods.go:61] "kube-apiserver-addons-382533" [7813eb75-ebef-455b-a3de-93b7a0643c2a] Running
I0910 17:31:22.873672 14014 system_pods.go:61] "kube-controller-manager-addons-382533" [97953725-3c5b-4c2b-9124-84281a223f29] Running
I0910 17:31:22.873677 14014 system_pods.go:61] "kube-ingress-dns-minikube" [7848fac9-1aa6-4c77-a9b5-f52257dd711a] Running
I0910 17:31:22.873680 14014 system_pods.go:61] "kube-proxy-29nvr" [017b1ee3-02c7-4c1b-8643-a15df1ae17a7] Running
I0910 17:31:22.873685 14014 system_pods.go:61] "kube-scheduler-addons-382533" [266946dc-53ff-4b4d-b8e7-a61faf948e95] Running
I0910 17:31:22.873689 14014 system_pods.go:61] "metrics-server-84c5f94fbc-p8qng" [ec83e369-bb84-4074-a414-917ab84c8b31] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0910 17:31:22.873694 14014 system_pods.go:61] "nvidia-device-plugin-daemonset-7bk7f" [59856e36-37bb-4399-bb83-d027a31b052a] Running
I0910 17:31:22.873698 14014 system_pods.go:61] "registry-66c9cd494c-zlfgb" [cb8eada4-03cf-4da1-8b47-5bce3b3d3226] Running
I0910 17:31:22.873704 14014 system_pods.go:61] "registry-proxy-jb2hn" [eccdebcf-344c-4fa8-9409-061fa7c8d2ab] Running
I0910 17:31:22.873710 14014 system_pods.go:61] "snapshot-controller-56fcc65765-6sjtt" [05399a10-ebbf-434a-91ba-802096e2d2b2] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0910 17:31:22.873718 14014 system_pods.go:61] "snapshot-controller-56fcc65765-whsxw" [6bbc521e-9ab6-46a8-9a23-87c9b8b0df78] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0910 17:31:22.873721 14014 system_pods.go:61] "storage-provisioner" [6083fb53-ea8b-4233-b1f4-d57a7cc0052d] Running
I0910 17:31:22.873725 14014 system_pods.go:61] "tiller-deploy-b48cc5f79-rpvpt" [f0312442-16e9-4e0e-89f3-3151b2d19bda] Running
I0910 17:31:22.873731 14014 system_pods.go:74] duration metric: took 187.510439ms to wait for pod list to return data ...
I0910 17:31:22.873742 14014 default_sa.go:34] waiting for default service account to be created ...
I0910 17:31:23.069639 14014 default_sa.go:45] found service account: "default"
I0910 17:31:23.069662 14014 default_sa.go:55] duration metric: took 195.912622ms for default service account to be created ...
I0910 17:31:23.069671 14014 system_pods.go:116] waiting for k8s-apps to be running ...
I0910 17:31:23.150099 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:23.163724 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:23.273997 14014 system_pods.go:86] 18 kube-system pods found
I0910 17:31:23.274027 14014 system_pods.go:89] "coredns-6f6b679f8f-g2x6g" [9ad2f9ec-61cf-4a29-a8ff-24e2bd3088db] Running
I0910 17:31:23.274036 14014 system_pods.go:89] "csi-hostpath-attacher-0" [06d0a0cc-d0f6-4e05-8af6-bf4f4d2d20fc] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0910 17:31:23.274042 14014 system_pods.go:89] "csi-hostpath-resizer-0" [dd5662db-6ebf-48a7-ad31-c70fad0c33fb] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0910 17:31:23.274050 14014 system_pods.go:89] "csi-hostpathplugin-9jbqt" [1b107483-67fb-46fb-8d25-9aa28471a833] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0910 17:31:23.274054 14014 system_pods.go:89] "etcd-addons-382533" [c444f86b-711d-4fc5-89eb-20cea8837776] Running
I0910 17:31:23.274059 14014 system_pods.go:89] "kube-apiserver-addons-382533" [7813eb75-ebef-455b-a3de-93b7a0643c2a] Running
I0910 17:31:23.274063 14014 system_pods.go:89] "kube-controller-manager-addons-382533" [97953725-3c5b-4c2b-9124-84281a223f29] Running
I0910 17:31:23.274067 14014 system_pods.go:89] "kube-ingress-dns-minikube" [7848fac9-1aa6-4c77-a9b5-f52257dd711a] Running
I0910 17:31:23.274070 14014 system_pods.go:89] "kube-proxy-29nvr" [017b1ee3-02c7-4c1b-8643-a15df1ae17a7] Running
I0910 17:31:23.274078 14014 system_pods.go:89] "kube-scheduler-addons-382533" [266946dc-53ff-4b4d-b8e7-a61faf948e95] Running
I0910 17:31:23.274087 14014 system_pods.go:89] "metrics-server-84c5f94fbc-p8qng" [ec83e369-bb84-4074-a414-917ab84c8b31] Running
I0910 17:31:23.274091 14014 system_pods.go:89] "nvidia-device-plugin-daemonset-7bk7f" [59856e36-37bb-4399-bb83-d027a31b052a] Running
I0910 17:31:23.274094 14014 system_pods.go:89] "registry-66c9cd494c-zlfgb" [cb8eada4-03cf-4da1-8b47-5bce3b3d3226] Running
I0910 17:31:23.274098 14014 system_pods.go:89] "registry-proxy-jb2hn" [eccdebcf-344c-4fa8-9409-061fa7c8d2ab] Running
I0910 17:31:23.274104 14014 system_pods.go:89] "snapshot-controller-56fcc65765-6sjtt" [05399a10-ebbf-434a-91ba-802096e2d2b2] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0910 17:31:23.274111 14014 system_pods.go:89] "snapshot-controller-56fcc65765-whsxw" [6bbc521e-9ab6-46a8-9a23-87c9b8b0df78] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0910 17:31:23.274116 14014 system_pods.go:89] "storage-provisioner" [6083fb53-ea8b-4233-b1f4-d57a7cc0052d] Running
I0910 17:31:23.274123 14014 system_pods.go:89] "tiller-deploy-b48cc5f79-rpvpt" [f0312442-16e9-4e0e-89f3-3151b2d19bda] Running
I0910 17:31:23.274129 14014 system_pods.go:126] duration metric: took 204.453334ms to wait for k8s-apps to be running ...
I0910 17:31:23.274138 14014 system_svc.go:44] waiting for kubelet service to be running ....
I0910 17:31:23.274181 14014 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0910 17:31:23.284951 14014 system_svc.go:56] duration metric: took 10.805601ms WaitForService to wait for kubelet
I0910 17:31:23.284979 14014 kubeadm.go:582] duration metric: took 44.57350813s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0910 17:31:23.285001 14014 node_conditions.go:102] verifying NodePressure condition ...
I0910 17:31:23.471387 14014 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0910 17:31:23.471417 14014 node_conditions.go:123] node cpu capacity is 8
I0910 17:31:23.471431 14014 node_conditions.go:105] duration metric: took 186.424312ms to run NodePressure ...
I0910 17:31:23.471444 14014 start.go:241] waiting for startup goroutines ...
I0910 17:31:23.650913 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:23.664981 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:24.149940 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:24.164378 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:24.649875 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:24.664034 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:25.149624 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:25.164277 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:25.649942 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:25.664018 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:26.150175 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:26.164450 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:26.649368 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:26.664161 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:27.149835 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:27.163757 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:27.649809 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:27.664447 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:28.150179 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:28.163901 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:28.649574 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:28.664629 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:29.293342 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:29.293348 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:29.650676 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:29.664154 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:30.149821 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:30.164614 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:30.650953 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:30.668051 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:31.149503 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:31.165885 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:31.649687 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:31.665651 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:32.149847 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:32.164445 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:32.650485 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:32.664637 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:33.150125 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:33.164635 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:33.650137 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:33.664126 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:34.149715 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:34.165021 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:34.650876 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:34.665248 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:35.149984 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:35.164845 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:35.650208 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:35.663690 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:36.149483 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:36.163980 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:36.649659 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:36.664467 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:37.149926 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:37.164810 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:37.649829 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:37.664751 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:38.149739 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:38.164158 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:38.650232 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:38.663815 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:39.149833 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:39.163545 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:39.649864 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:39.663321 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:40.150186 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:40.164430 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:40.650051 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:40.664589 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:41.149795 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:41.164674 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:41.650979 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:41.664767 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:42.150007 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:42.164306 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:42.650601 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:42.664573 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:43.150345 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:43.163887 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:43.650768 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:43.664767 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:44.150014 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:44.163748 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:44.650534 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:44.664194 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:45.150000 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:45.164279 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:45.652312 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:45.663747 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:46.149964 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:46.163942 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:46.649360 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:46.750301 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:47.150049 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:47.164518 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:47.650184 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:47.666556 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:48.149748 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:48.164990 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:48.649636 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:48.665014 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:49.150040 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:49.163692 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:49.649637 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:49.664459 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0910 17:31:50.149518 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:50.164158 14014 kapi.go:107] duration metric: took 57.504113765s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0910 17:31:50.649174 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:51.149666 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:51.650325 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:52.149956 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:52.649991 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:53.150656 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:53.650744 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:54.149702 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:54.649533 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:55.150968 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:55.649307 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:56.150307 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:56.649899 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:57.149645 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:57.649993 14014 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0910 17:31:58.150239 14014 kapi.go:107] duration metric: took 1m7.004635175s to wait for app.kubernetes.io/name=ingress-nginx ...
I0910 17:32:17.681167 14014 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0910 17:32:17.681188 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:18.181241 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:18.681407 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:19.181656 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:19.681419 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:20.181710 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:20.682061 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:21.181027 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:21.681309 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:22.181308 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:22.681881 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:23.181327 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:23.681208 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:24.181336 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:24.681072 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:25.181635 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:25.681518 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:26.181454 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:26.681449 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:27.181490 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:27.681738 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:28.181668 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:28.681670 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:29.182030 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:29.680873 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:30.181945 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:30.680779 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:31.182025 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:31.680891 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:32.181794 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:32.682088 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:33.181139 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:33.681261 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:34.181316 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:34.681228 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:35.181546 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:35.681675 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:36.181746 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:36.681942 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:37.181347 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:37.680976 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:38.181175 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:38.680915 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:39.180992 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:39.680748 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:40.181084 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:40.680936 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:41.180964 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:41.681606 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:42.181745 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:42.681803 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:43.181444 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:43.681169 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:44.181376 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:44.682287 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:45.181383 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:45.681301 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:46.181880 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:46.681220 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:47.180984 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:47.680862 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:48.181735 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:48.681402 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:49.181891 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:49.681543 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:50.181810 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:50.681537 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:51.181651 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:51.681949 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:52.181152 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:52.681418 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:53.181813 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:53.681656 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:54.181810 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:54.681897 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:55.181148 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:55.680999 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:56.181119 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:56.680835 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:57.181181 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:57.681040 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:58.181251 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:58.680994 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:59.196532 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:32:59.680599 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:00.181481 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:00.681349 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:01.181037 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:01.680950 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:02.181011 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:02.681166 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:03.181117 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:03.680929 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:04.180991 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:04.681356 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:05.181717 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:05.681475 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:06.181569 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:06.681894 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:07.181003 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:07.680867 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:08.181645 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:08.681508 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:09.181318 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:09.681281 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:10.181565 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:10.681529 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:11.181912 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:11.680955 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:12.181218 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:12.681040 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:13.180723 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:13.681366 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:14.181992 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:14.680860 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:15.181367 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:15.680991 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:16.181125 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:16.681026 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:17.181757 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:17.681582 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:18.181770 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:18.681442 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:19.181397 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:19.680911 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:20.181021 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:20.680596 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:21.181740 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:21.682024 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:22.180997 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:22.681021 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:23.180972 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:23.680843 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:24.182009 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:24.681275 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:25.184249 14014 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0910 17:33:25.681489 14014 kapi.go:107] duration metric: took 2m31.503486272s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0910 17:33:25.683155 14014 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-382533 cluster.
I0910 17:33:25.684408 14014 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0910 17:33:25.685502 14014 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0910 17:33:25.686897 14014 out.go:177] * Enabled addons: storage-provisioner, ingress-dns, nvidia-device-plugin, cloud-spanner, default-storageclass, volcano, helm-tiller, metrics-server, yakd, inspektor-gadget, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0910 17:33:25.688039 14014 addons.go:510] duration metric: took 2m46.976507249s for enable addons: enabled=[storage-provisioner ingress-dns nvidia-device-plugin cloud-spanner default-storageclass volcano helm-tiller metrics-server yakd inspektor-gadget volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0910 17:33:25.688077 14014 start.go:246] waiting for cluster config update ...
I0910 17:33:25.688106 14014 start.go:255] writing updated cluster config ...
I0910 17:33:25.688382 14014 ssh_runner.go:195] Run: rm -f paused
I0910 17:33:25.735845 14014 start.go:600] kubectl: 1.31.0, cluster: 1.31.0 (minor skew: 0)
I0910 17:33:25.737441 14014 out.go:177] * Done! kubectl is now configured to use "addons-382533" cluster and "default" namespace by default
==> Docker <==
Sep 10 17:42:54 addons-382533 dockerd[1341]: time="2024-09-10T17:42:54.247273321Z" level=info msg="ignoring event" container=34f6b7bc4c6b51e9a1145a691411c116283e5a1f218acaa9efb3b3bf24d890d0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:42:54 addons-382533 dockerd[1341]: time="2024-09-10T17:42:54.286983144Z" level=info msg="ignoring event" container=f11e3df4b7924641dc6d6848a943e0bb20b96278bd541a02f3691722e68c5cde module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:42:54 addons-382533 dockerd[1341]: time="2024-09-10T17:42:54.306530187Z" level=info msg="ignoring event" container=230bbca04659cff38e3fc8db846b7528933eb468a29785690bfe45c1083d8248 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:42:56 addons-382533 dockerd[1341]: time="2024-09-10T17:42:56.480563319Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 10 17:42:56 addons-382533 dockerd[1341]: time="2024-09-10T17:42:56.482593206Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 10 17:42:58 addons-382533 cri-dockerd[1606]: time="2024-09-10T17:42:58Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/fdb943312d8133a9b9fb344da5979b654f8f005335d03e077df048b214ed4b3e/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local europe-west4-a.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 10 17:42:58 addons-382533 dockerd[1341]: time="2024-09-10T17:42:58.460231571Z" level=info msg="ignoring event" container=e727a5709396eb96eef06b944a064c7ff0d1512d6afdc47cf9c1f55b090666a2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:42:58 addons-382533 dockerd[1341]: time="2024-09-10T17:42:58.505944351Z" level=info msg="ignoring event" container=ae86712dfbc7f75c17e7b43017b78298d27f3618a4e6e99b7c030f0905df61d2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:00 addons-382533 cri-dockerd[1606]: time="2024-09-10T17:43:00Z" level=info msg="Stop pulling image docker.io/kicbase/echo-server:1.0: Status: Downloaded newer image for kicbase/echo-server:1.0"
Sep 10 17:43:00 addons-382533 dockerd[1341]: time="2024-09-10T17:43:00.366348622Z" level=info msg="ignoring event" container=210b3a188baedf00c3a57e2f70f153c57c883a9bcb2cb6c137dbba9ded44feb0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:00 addons-382533 dockerd[1341]: time="2024-09-10T17:43:00.369281595Z" level=info msg="ignoring event" container=12e554eabd1605df2b5327ffb9f30f3b35930118d0c8c8e7c6f39735aba18123 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:00 addons-382533 dockerd[1341]: time="2024-09-10T17:43:00.572485031Z" level=info msg="ignoring event" container=4cb2e76096fdd963d8db61a116012f4ea6c59af8fc5d2c26360082ce105a800e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:00 addons-382533 dockerd[1341]: time="2024-09-10T17:43:00.595445719Z" level=info msg="ignoring event" container=f561a7d8dffcc352b2d4a007e3881b4c9a39880d94f7e00bc2a6ba2a6a84c73e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:02 addons-382533 dockerd[1341]: time="2024-09-10T17:43:02.699291133Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=38b10c7a48109b60c692bca6c09811545fe8970b879c80f824fc45e33fd415c3
Sep 10 17:43:02 addons-382533 dockerd[1341]: time="2024-09-10T17:43:02.752902448Z" level=info msg="ignoring event" container=38b10c7a48109b60c692bca6c09811545fe8970b879c80f824fc45e33fd415c3 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:02 addons-382533 dockerd[1341]: time="2024-09-10T17:43:02.895028314Z" level=info msg="ignoring event" container=9817732be6c316d726e3262360b7d95686574830916691405f32fb1fa1c3a159 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:03 addons-382533 cri-dockerd[1606]: time="2024-09-10T17:43:03Z" level=error msg="error getting RW layer size for container ID 'e727a5709396eb96eef06b944a064c7ff0d1512d6afdc47cf9c1f55b090666a2': Error response from daemon: No such container: e727a5709396eb96eef06b944a064c7ff0d1512d6afdc47cf9c1f55b090666a2"
Sep 10 17:43:03 addons-382533 cri-dockerd[1606]: time="2024-09-10T17:43:03Z" level=error msg="Set backoffDuration to : 1m0s for container ID 'e727a5709396eb96eef06b944a064c7ff0d1512d6afdc47cf9c1f55b090666a2'"
Sep 10 17:43:04 addons-382533 cri-dockerd[1606]: time="2024-09-10T17:43:04Z" level=error msg="error getting RW layer size for container ID '38b10c7a48109b60c692bca6c09811545fe8970b879c80f824fc45e33fd415c3': Error response from daemon: No such container: 38b10c7a48109b60c692bca6c09811545fe8970b879c80f824fc45e33fd415c3"
Sep 10 17:43:04 addons-382533 cri-dockerd[1606]: time="2024-09-10T17:43:04Z" level=error msg="Set backoffDuration to : 1m0s for container ID '38b10c7a48109b60c692bca6c09811545fe8970b879c80f824fc45e33fd415c3'"
Sep 10 17:43:19 addons-382533 dockerd[1341]: time="2024-09-10T17:43:19.032143645Z" level=info msg="ignoring event" container=418ea104c08cfc68c6fc57ceb8797668f2d3675662c89e7177432149cd947b5f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:19 addons-382533 dockerd[1341]: time="2024-09-10T17:43:19.543449445Z" level=info msg="ignoring event" container=f8efe6be42bea8aa796f7af40f816ce95ea1bb2ddbe464f4b13f11fefeb1707a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:19 addons-382533 dockerd[1341]: time="2024-09-10T17:43:19.557808646Z" level=info msg="ignoring event" container=98a126179d6fbc80c0def47519d7a7a3fad4e049faa72c54b859221cf8708f79 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:19 addons-382533 dockerd[1341]: time="2024-09-10T17:43:19.751041115Z" level=info msg="ignoring event" container=c609e55c2520dc83d53a5f215cf185da5538577082475b3c737df361851d0067 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 10 17:43:19 addons-382533 dockerd[1341]: time="2024-09-10T17:43:19.775186278Z" level=info msg="ignoring event" container=3c9ce84e8d0ae3f1f8e21b6161547abe132bb301f0fcdba12e8722e6b8136cc6 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
0aa41d7106552 kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 20 seconds ago Running hello-world-app 0 fdb943312d813 hello-world-app-55bf9c44b4-sl2ht
e7504ca1b9abe nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf 29 seconds ago Running nginx 0 40255d21f4988 nginx
8ab8bcdefa911 busybox@sha256:34b191d63fbc93e25e275bfccf1b5365664e5ac28f06d974e8d50090fbb49f41 44 seconds ago Exited busybox 0 42bfe0cf21310 test-local-path
351d362f9e541 gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 8916e20daefaa gcp-auth-89d5ffd79-qtklr
763c784738143 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited patch 0 2590846b6e5e7 ingress-nginx-admission-patch-8f4pg
b8b1f0b1d66fb registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 f48fd70132af9 ingress-nginx-admission-create-cxtbk
f26f8d9ddb4a7 rancher/local-path-provisioner@sha256:e34c88ae0affb1cdefbb874140d6339d4a27ec4ee420ae8199cd839997b05246 12 minutes ago Running local-path-provisioner 0 1c00821d3d54e local-path-provisioner-86d989889c-fr6qj
98a126179d6fb gcr.io/k8s-minikube/kube-registry-proxy@sha256:b3fa0b2df8737fdb85ad5918a7e2652527463e357afff83a5e5bb966bcedc367 12 minutes ago Exited registry-proxy 0 3c9ce84e8d0ae registry-proxy-jb2hn
f8efe6be42bea registry@sha256:ac0192b549007e22998eb74e8d8488dcfe70f1489520c3b144a6047ac5efbe90 12 minutes ago Exited registry 0 c609e55c2520d registry-66c9cd494c-zlfgb
858ca178e3667 6e38f40d628db 12 minutes ago Running storage-provisioner 0 3543fc8973218 storage-provisioner
e97966ccd2fa0 cbb01a7bd410d 12 minutes ago Running coredns 0 a7592530d7f79 coredns-6f6b679f8f-g2x6g
f447ac50f3f27 ad83b2ca7b09e 12 minutes ago Running kube-proxy 0 d33084f70e9a2 kube-proxy-29nvr
fc132bb251256 1766f54c897f0 12 minutes ago Running kube-scheduler 0 ec9341613da29 kube-scheduler-addons-382533
04b1b3c8332cd 604f5db92eaa8 12 minutes ago Running kube-apiserver 0 f1a04d3bd8d82 kube-apiserver-addons-382533
927a52be42b90 2e96e5913fc06 12 minutes ago Running etcd 0 62e8f69ea80a0 etcd-addons-382533
be82cec3a1671 045733566833c 12 minutes ago Running kube-controller-manager 0 7a45731f27749 kube-controller-manager-addons-382533
==> coredns [e97966ccd2fa] <==
[INFO] 10.244.0.22:43034 - 23579 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000185324s
[INFO] 10.244.0.22:37970 - 11479 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005216183s
[INFO] 10.244.0.22:39046 - 30610 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.006014182s
[INFO] 10.244.0.22:57058 - 11508 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.002438298s
[INFO] 10.244.0.22:37438 - 16461 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000382113s
[INFO] 10.244.0.22:50296 - 14633 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000379843s
[INFO] 10.244.0.22:44680 - 31750 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000488978s
[INFO] 10.244.0.22:39046 - 31913 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000498392s
[INFO] 10.244.0.22:54879 - 53172 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000489993s
[INFO] 10.244.0.22:39176 - 37546 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000062058s
[INFO] 10.244.0.22:37715 - 1126 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000540238s
[INFO] 10.244.0.22:60297 - 34113 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000573322s
[INFO] 10.244.0.22:57058 - 52433 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000648926s
[INFO] 10.244.0.22:37970 - 6202 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000167801s
[INFO] 10.244.0.22:51847 - 22214 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.004263738s
[INFO] 10.244.0.22:59744 - 12973 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.004028861s
[INFO] 10.244.0.22:54539 - 27879 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003588302s
[INFO] 10.244.0.22:51847 - 65165 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004389294s
[INFO] 10.244.0.22:59744 - 3869 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004190922s
[INFO] 10.244.0.22:54539 - 20464 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005116725s
[INFO] 10.244.0.22:51847 - 2111 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.001342299s
[INFO] 10.244.0.22:51847 - 15277 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000082323s
[INFO] 10.244.0.22:54539 - 20909 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00004228s
[INFO] 10.244.0.22:59744 - 31223 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.012699094s
[INFO] 10.244.0.22:59744 - 12920 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000072297s
==> describe nodes <==
Name: addons-382533
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-382533
kubernetes.io/os=linux
minikube.k8s.io/commit=37b4bace07cd53444288cad630e4db4b688b8c18
minikube.k8s.io/name=addons-382533
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_10T17_30_34_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-382533
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Tue, 10 Sep 2024 17:30:31 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-382533
AcquireTime: <unset>
RenewTime: Tue, 10 Sep 2024 17:43:19 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Tue, 10 Sep 2024 17:43:09 +0000 Tue, 10 Sep 2024 17:30:29 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Tue, 10 Sep 2024 17:43:09 +0000 Tue, 10 Sep 2024 17:30:29 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Tue, 10 Sep 2024 17:43:09 +0000 Tue, 10 Sep 2024 17:30:29 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Tue, 10 Sep 2024 17:43:09 +0000 Tue, 10 Sep 2024 17:30:31 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-382533
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859320Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859320Ki
pods: 110
System Info:
Machine ID: c319f92c245b468ea1536b2d8684acc7
System UUID: 754c8113-2ec9-425f-b593-d494a5bbaf7b
Boot ID: a4d7b1d8-0129-48fb-9a81-dfef4e5f0f2f
Kernel Version: 5.15.0-1067-gcp
OS Image: Ubuntu 22.04.4 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.2.1
Kubelet Version: v1.31.0
Kube-Proxy Version:
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (12 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m14s
default hello-world-app-55bf9c44b4-sl2ht 0 (0%) 0 (0%) 0 (0%) 0 (0%) 23s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 33s
gcp-auth gcp-auth-89d5ffd79-qtklr 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
kube-system coredns-6f6b679f8f-g2x6g 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-382533 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-382533 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-382533 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-29nvr 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-382533 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
local-path-storage local-path-provisioner-86d989889c-fr6qj 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeHasSufficientMemory 12m (x8 over 12m) kubelet Node addons-382533 status is now: NodeHasSufficientMemory
Normal NodeHasSufficientPID 12m (x7 over 12m) kubelet Node addons-382533 status is now: NodeHasSufficientPID
Normal NodeHasNoDiskPressure 12m (x7 over 12m) kubelet Node addons-382533 status is now: NodeHasNoDiskPressure
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-382533 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-382533 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-382533 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-382533 event: Registered Node addons-382533 in Controller
==> dmesg <==
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff ae f2 2f 07 3a 94 08 06
[ +0.127672] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff fa b2 54 6b 9a 88 08 06
[ +0.021115] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 4a 1b 0b 7f 22 8b 08 06
[ +10.335529] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 4e 03 7e df b3 23 08 06
[ +1.062609] IPv4: martian source 10.244.0.1 from 10.244.0.23, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 92 7e 48 c0 70 d8 08 06
[Sep10 17:32] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 12 fc a2 22 52 c9 08 06
[ +0.071236] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 0a 79 9e 29 6b 93 08 06
[Sep10 17:33] IPv4: martian source 10.244.0.1 from 10.244.0.26, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 02 cc b6 a5 e9 55 08 06
[ +0.000484] IPv4: martian source 10.244.0.26 from 10.244.0.3, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff e2 5b 28 5f 44 ac 08 06
[Sep10 17:42] IPv4: martian source 10.244.0.1 from 10.244.0.29, on dev eth0
[ +0.000010] ll header: 00000000: ff ff ff ff ff ff 02 35 8a ad c8 c3 08 06
[ +3.514600] IPv4: martian source 10.244.0.1 from 10.244.0.31, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 52 70 94 4b 6d e5 08 06
[ +37.209021] IPv4: martian source 10.244.0.38 from 10.244.0.22, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 4e 03 7e df b3 23 08 06
[ +1.290345] IPv4: martian source 10.244.0.22 from 10.244.0.3, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff e2 5b 28 5f 44 ac 08 06
==> etcd [927a52be42b9] <==
{"level":"info","ts":"2024-09-10T17:30:29.746262Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-10T17:30:29.746337Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-10T17:30:29.746358Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-10T17:30:29.746376Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-10T17:30:29.746387Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-10T17:30:29.747076Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-10T17:30:29.747243Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-10T17:30:29.747910Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-09-10T17:30:29.748036Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-10T17:30:48.959643Z","caller":"traceutil/trace.go:171","msg":"trace[232656297] transaction","detail":"{read_only:false; response_revision:770; number_of_response:1; }","duration":"106.514968ms","start":"2024-09-10T17:30:48.853109Z","end":"2024-09-10T17:30:48.959624Z","steps":["trace[232656297] 'process raft request' (duration: 99.829297ms)"],"step_count":1}
{"level":"info","ts":"2024-09-10T17:30:48.959994Z","caller":"traceutil/trace.go:171","msg":"trace[1446142945] transaction","detail":"{read_only:false; response_revision:771; number_of_response:1; }","duration":"100.191658ms","start":"2024-09-10T17:30:48.859792Z","end":"2024-09-10T17:30:48.959984Z","steps":["trace[1446142945] 'process raft request' (duration: 99.412182ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-10T17:30:48.960842Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"100.065061ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/deployments/volcano-system/volcano-admission\" ","response":"range_response_count:1 size:5238"}
{"level":"info","ts":"2024-09-10T17:30:48.960895Z","caller":"traceutil/trace.go:171","msg":"trace[1139493827] range","detail":"{range_begin:/registry/deployments/volcano-system/volcano-admission; range_end:; response_count:1; response_revision:772; }","duration":"101.099893ms","start":"2024-09-10T17:30:48.859782Z","end":"2024-09-10T17:30:48.960882Z","steps":["trace[1139493827] 'agreement among raft nodes before linearized reading' (duration: 100.004421ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-10T17:31:11.120229Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"104.189104ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/services/endpoints/kube-system/k8s.io-minikube-hostpath\" ","response":"range_response_count:1 size:1113"}
{"level":"info","ts":"2024-09-10T17:31:11.120294Z","caller":"traceutil/trace.go:171","msg":"trace[1637018196] range","detail":"{range_begin:/registry/services/endpoints/kube-system/k8s.io-minikube-hostpath; range_end:; response_count:1; response_revision:1037; }","duration":"104.26328ms","start":"2024-09-10T17:31:11.016016Z","end":"2024-09-10T17:31:11.120279Z","steps":["trace[1637018196] 'range keys from in-memory index tree' (duration: 104.006865ms)"],"step_count":1}
{"level":"info","ts":"2024-09-10T17:31:11.238083Z","caller":"traceutil/trace.go:171","msg":"trace[1031508079] transaction","detail":"{read_only:false; response_revision:1038; number_of_response:1; }","duration":"115.004583ms","start":"2024-09-10T17:31:11.123053Z","end":"2024-09-10T17:31:11.238058Z","steps":["trace[1031508079] 'process raft request' (duration: 114.879403ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-10T17:31:29.290369Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"110.605197ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"warn","ts":"2024-09-10T17:31:29.290381Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"143.025999ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-10T17:31:29.290438Z","caller":"traceutil/trace.go:171","msg":"trace[138576467] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:1126; }","duration":"110.69466ms","start":"2024-09-10T17:31:29.179729Z","end":"2024-09-10T17:31:29.290424Z","steps":["trace[138576467] 'range keys from in-memory index tree' (duration: 110.549666ms)"],"step_count":1}
{"level":"info","ts":"2024-09-10T17:31:29.290442Z","caller":"traceutil/trace.go:171","msg":"trace[2025045550] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:1126; }","duration":"143.094505ms","start":"2024-09-10T17:31:29.147336Z","end":"2024-09-10T17:31:29.290430Z","steps":["trace[2025045550] 'range keys from in-memory index tree' (duration: 142.96853ms)"],"step_count":1}
{"level":"warn","ts":"2024-09-10T17:31:29.290524Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"128.777481ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-10T17:31:29.290562Z","caller":"traceutil/trace.go:171","msg":"trace[841850304] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:1126; }","duration":"128.819038ms","start":"2024-09-10T17:31:29.161732Z","end":"2024-09-10T17:31:29.290551Z","steps":["trace[841850304] 'range keys from in-memory index tree' (duration: 128.722331ms)"],"step_count":1}
{"level":"info","ts":"2024-09-10T17:40:29.771107Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1914}
{"level":"info","ts":"2024-09-10T17:40:29.796305Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1914,"took":"24.681874ms","hash":4186327218,"current-db-size-bytes":9252864,"current-db-size":"9.3 MB","current-db-size-in-use-bytes":5148672,"current-db-size-in-use":"5.1 MB"}
{"level":"info","ts":"2024-09-10T17:40:29.796346Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":4186327218,"revision":1914,"compact-revision":-1}
==> gcp-auth [351d362f9e54] <==
2024/09/10 17:42:13 Ready to write response ...
2024/09/10 17:42:18 Ready to marshal response ...
2024/09/10 17:42:18 Ready to write response ...
2024/09/10 17:42:19 Ready to marshal response ...
2024/09/10 17:42:19 Ready to write response ...
2024/09/10 17:42:25 Ready to marshal response ...
2024/09/10 17:42:25 Ready to write response ...
2024/09/10 17:42:28 Ready to marshal response ...
2024/09/10 17:42:28 Ready to write response ...
2024/09/10 17:42:28 Ready to marshal response ...
2024/09/10 17:42:28 Ready to write response ...
2024/09/10 17:42:36 Ready to marshal response ...
2024/09/10 17:42:36 Ready to write response ...
2024/09/10 17:42:36 Ready to marshal response ...
2024/09/10 17:42:36 Ready to write response ...
2024/09/10 17:42:36 Ready to marshal response ...
2024/09/10 17:42:36 Ready to write response ...
2024/09/10 17:42:39 Ready to marshal response ...
2024/09/10 17:42:39 Ready to write response ...
2024/09/10 17:42:44 Ready to marshal response ...
2024/09/10 17:42:44 Ready to write response ...
2024/09/10 17:42:47 Ready to marshal response ...
2024/09/10 17:42:47 Ready to write response ...
2024/09/10 17:42:57 Ready to marshal response ...
2024/09/10 17:42:57 Ready to write response ...
==> kernel <==
17:43:20 up 25 min, 0 users, load average: 0.64, 0.52, 0.49
Linux addons-382533 5.15.0-1067-gcp #75~20.04.1-Ubuntu SMP Wed Aug 7 20:43:22 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.4 LTS"
==> kube-apiserver [04b1b3c8332c] <==
W0910 17:33:58.474620 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0910 17:33:58.776689 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0910 17:33:59.074589 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0910 17:42:14.191183 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0910 17:42:15.206897 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
E0910 17:42:16.878304 1 conn.go:339] Error on socket receive: read tcp 192.168.49.2:8443->192.168.49.1:38856: use of closed network connection
E0910 17:42:20.396742 1 upgradeaware.go:427] Error proxying data from client to backend: read tcp 192.168.49.2:8443->10.244.0.31:40908: read: connection reset by peer
I0910 17:42:34.141230 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0910 17:42:36.112002 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.109.71.43"}
I0910 17:42:47.157428 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0910 17:42:47.323128 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.105.229.135"}
I0910 17:42:57.799347 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.105.194.68"}
I0910 17:43:00.143321 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0910 17:43:00.143379 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0910 17:43:00.161364 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0910 17:43:00.161406 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0910 17:43:00.161904 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0910 17:43:00.161934 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0910 17:43:00.170543 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0910 17:43:00.170591 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0910 17:43:00.255287 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0910 17:43:00.255328 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0910 17:43:01.162442 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W0910 17:43:01.255728 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
W0910 17:43:01.345887 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
==> kube-controller-manager [be82cec3a167] <==
W0910 17:43:04.517544 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0910 17:43:04.517588 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0910 17:43:04.807507 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0910 17:43:04.807546 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0910 17:43:06.074715 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0910 17:43:06.074754 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0910 17:43:08.351733 1 shared_informer.go:313] Waiting for caches to sync for resource quota
I0910 17:43:08.351772 1 shared_informer.go:320] Caches are synced for resource quota
I0910 17:43:08.650943 1 shared_informer.go:313] Waiting for caches to sync for garbage collector
I0910 17:43:08.650995 1 shared_informer.go:320] Caches are synced for garbage collector
W0910 17:43:08.780025 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0910 17:43:08.780063 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0910 17:43:08.919249 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0910 17:43:08.919289 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0910 17:43:09.440930 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-382533"
I0910 17:43:09.740122 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="ingress-nginx"
W0910 17:43:10.433954 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0910 17:43:10.433990 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0910 17:43:12.616891 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0910 17:43:12.616927 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0910 17:43:16.749301 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0910 17:43:16.749342 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0910 17:43:19.459802 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="6.522µs"
W0910 17:43:20.502666 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0910 17:43:20.502713 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
==> kube-proxy [f447ac50f3f2] <==
I0910 17:30:41.343114 1 server_linux.go:66] "Using iptables proxy"
I0910 17:30:41.943505 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0910 17:30:41.943581 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0910 17:30:42.260947 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0910 17:30:42.261006 1 server_linux.go:169] "Using iptables Proxier"
I0910 17:30:42.264305 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0910 17:30:42.264771 1 server.go:483] "Version info" version="v1.31.0"
I0910 17:30:42.264795 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0910 17:30:42.267668 1 config.go:326] "Starting node config controller"
I0910 17:30:42.267689 1 shared_informer.go:313] Waiting for caches to sync for node config
I0910 17:30:42.267959 1 config.go:197] "Starting service config controller"
I0910 17:30:42.267964 1 shared_informer.go:313] Waiting for caches to sync for service config
I0910 17:30:42.267975 1 config.go:104] "Starting endpoint slice config controller"
I0910 17:30:42.267978 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0910 17:30:42.446125 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0910 17:30:42.446194 1 shared_informer.go:320] Caches are synced for node config
I0910 17:30:42.446205 1 shared_informer.go:320] Caches are synced for service config
==> kube-scheduler [fc132bb25125] <==
W0910 17:30:31.156226 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0910 17:30:31.156455 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0910 17:30:31.156477 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
W0910 17:30:31.156228 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0910 17:30:31.156494 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
E0910 17:30:31.156501 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0910 17:30:31.156230 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0910 17:30:31.156537 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0910 17:30:31.967535 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0910 17:30:31.967578 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0910 17:30:32.063225 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0910 17:30:32.063281 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0910 17:30:32.072573 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0910 17:30:32.072604 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0910 17:30:32.143840 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0910 17:30:32.143903 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0910 17:30:32.143843 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0910 17:30:32.143955 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0910 17:30:32.171478 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0910 17:30:32.171534 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0910 17:30:32.197805 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0910 17:30:32.197850 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0910 17:30:32.223144 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0910 17:30:32.223186 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
I0910 17:30:32.454513 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 10 17:43:03 addons-382533 kubelet[2457]: E0910 17:43:03.339971 2457 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 38b10c7a48109b60c692bca6c09811545fe8970b879c80f824fc45e33fd415c3" containerID="38b10c7a48109b60c692bca6c09811545fe8970b879c80f824fc45e33fd415c3"
Sep 10 17:43:03 addons-382533 kubelet[2457]: I0910 17:43:03.340008 2457 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"38b10c7a48109b60c692bca6c09811545fe8970b879c80f824fc45e33fd415c3"} err="failed to get container status \"38b10c7a48109b60c692bca6c09811545fe8970b879c80f824fc45e33fd415c3\": rpc error: code = Unknown desc = Error response from daemon: No such container: 38b10c7a48109b60c692bca6c09811545fe8970b879c80f824fc45e33fd415c3"
Sep 10 17:43:03 addons-382533 kubelet[2457]: I0910 17:43:03.357900 2457 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea8820ed-f1aa-45a5-a47a-41404465bcd5" path="/var/lib/kubelet/pods/ea8820ed-f1aa-45a5-a47a-41404465bcd5/volumes"
Sep 10 17:43:08 addons-382533 kubelet[2457]: E0910 17:43:08.351637 2457 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="03d365a2-b3d1-40f1-b1dc-34b9bc55a823"
Sep 10 17:43:11 addons-382533 kubelet[2457]: E0910 17:43:11.351103 2457 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox\\\"\"" pod="default/registry-test" podUID="2154fc57-0e79-4b33-94cd-8b9a95d5d63d"
Sep 10 17:43:19 addons-382533 kubelet[2457]: I0910 17:43:19.193468 2457 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvbpp\" (UniqueName: \"kubernetes.io/projected/2154fc57-0e79-4b33-94cd-8b9a95d5d63d-kube-api-access-bvbpp\") pod \"2154fc57-0e79-4b33-94cd-8b9a95d5d63d\" (UID: \"2154fc57-0e79-4b33-94cd-8b9a95d5d63d\") "
Sep 10 17:43:19 addons-382533 kubelet[2457]: I0910 17:43:19.193517 2457 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/2154fc57-0e79-4b33-94cd-8b9a95d5d63d-gcp-creds\") pod \"2154fc57-0e79-4b33-94cd-8b9a95d5d63d\" (UID: \"2154fc57-0e79-4b33-94cd-8b9a95d5d63d\") "
Sep 10 17:43:19 addons-382533 kubelet[2457]: I0910 17:43:19.193600 2457 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2154fc57-0e79-4b33-94cd-8b9a95d5d63d-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "2154fc57-0e79-4b33-94cd-8b9a95d5d63d" (UID: "2154fc57-0e79-4b33-94cd-8b9a95d5d63d"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Sep 10 17:43:19 addons-382533 kubelet[2457]: I0910 17:43:19.195335 2457 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2154fc57-0e79-4b33-94cd-8b9a95d5d63d-kube-api-access-bvbpp" (OuterVolumeSpecName: "kube-api-access-bvbpp") pod "2154fc57-0e79-4b33-94cd-8b9a95d5d63d" (UID: "2154fc57-0e79-4b33-94cd-8b9a95d5d63d"). InnerVolumeSpecName "kube-api-access-bvbpp". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 10 17:43:19 addons-382533 kubelet[2457]: I0910 17:43:19.293707 2457 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/2154fc57-0e79-4b33-94cd-8b9a95d5d63d-gcp-creds\") on node \"addons-382533\" DevicePath \"\""
Sep 10 17:43:19 addons-382533 kubelet[2457]: I0910 17:43:19.293736 2457 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-bvbpp\" (UniqueName: \"kubernetes.io/projected/2154fc57-0e79-4b33-94cd-8b9a95d5d63d-kube-api-access-bvbpp\") on node \"addons-382533\" DevicePath \"\""
Sep 10 17:43:19 addons-382533 kubelet[2457]: I0910 17:43:19.945918 2457 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d78sn\" (UniqueName: \"kubernetes.io/projected/eccdebcf-344c-4fa8-9409-061fa7c8d2ab-kube-api-access-d78sn\") pod \"eccdebcf-344c-4fa8-9409-061fa7c8d2ab\" (UID: \"eccdebcf-344c-4fa8-9409-061fa7c8d2ab\") "
Sep 10 17:43:19 addons-382533 kubelet[2457]: I0910 17:43:19.945981 2457 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghsk9\" (UniqueName: \"kubernetes.io/projected/cb8eada4-03cf-4da1-8b47-5bce3b3d3226-kube-api-access-ghsk9\") pod \"cb8eada4-03cf-4da1-8b47-5bce3b3d3226\" (UID: \"cb8eada4-03cf-4da1-8b47-5bce3b3d3226\") "
Sep 10 17:43:19 addons-382533 kubelet[2457]: I0910 17:43:19.948376 2457 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb8eada4-03cf-4da1-8b47-5bce3b3d3226-kube-api-access-ghsk9" (OuterVolumeSpecName: "kube-api-access-ghsk9") pod "cb8eada4-03cf-4da1-8b47-5bce3b3d3226" (UID: "cb8eada4-03cf-4da1-8b47-5bce3b3d3226"). InnerVolumeSpecName "kube-api-access-ghsk9". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 10 17:43:19 addons-382533 kubelet[2457]: I0910 17:43:19.948401 2457 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eccdebcf-344c-4fa8-9409-061fa7c8d2ab-kube-api-access-d78sn" (OuterVolumeSpecName: "kube-api-access-d78sn") pod "eccdebcf-344c-4fa8-9409-061fa7c8d2ab" (UID: "eccdebcf-344c-4fa8-9409-061fa7c8d2ab"). InnerVolumeSpecName "kube-api-access-d78sn". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 10 17:43:20 addons-382533 kubelet[2457]: I0910 17:43:20.046394 2457 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-d78sn\" (UniqueName: \"kubernetes.io/projected/eccdebcf-344c-4fa8-9409-061fa7c8d2ab-kube-api-access-d78sn\") on node \"addons-382533\" DevicePath \"\""
Sep 10 17:43:20 addons-382533 kubelet[2457]: I0910 17:43:20.046431 2457 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-ghsk9\" (UniqueName: \"kubernetes.io/projected/cb8eada4-03cf-4da1-8b47-5bce3b3d3226-kube-api-access-ghsk9\") on node \"addons-382533\" DevicePath \"\""
Sep 10 17:43:20 addons-382533 kubelet[2457]: I0910 17:43:20.510486 2457 scope.go:117] "RemoveContainer" containerID="f8efe6be42bea8aa796f7af40f816ce95ea1bb2ddbe464f4b13f11fefeb1707a"
Sep 10 17:43:20 addons-382533 kubelet[2457]: I0910 17:43:20.526649 2457 scope.go:117] "RemoveContainer" containerID="f8efe6be42bea8aa796f7af40f816ce95ea1bb2ddbe464f4b13f11fefeb1707a"
Sep 10 17:43:20 addons-382533 kubelet[2457]: E0910 17:43:20.527517 2457 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: f8efe6be42bea8aa796f7af40f816ce95ea1bb2ddbe464f4b13f11fefeb1707a" containerID="f8efe6be42bea8aa796f7af40f816ce95ea1bb2ddbe464f4b13f11fefeb1707a"
Sep 10 17:43:20 addons-382533 kubelet[2457]: I0910 17:43:20.527561 2457 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"f8efe6be42bea8aa796f7af40f816ce95ea1bb2ddbe464f4b13f11fefeb1707a"} err="failed to get container status \"f8efe6be42bea8aa796f7af40f816ce95ea1bb2ddbe464f4b13f11fefeb1707a\": rpc error: code = Unknown desc = Error response from daemon: No such container: f8efe6be42bea8aa796f7af40f816ce95ea1bb2ddbe464f4b13f11fefeb1707a"
Sep 10 17:43:20 addons-382533 kubelet[2457]: I0910 17:43:20.527590 2457 scope.go:117] "RemoveContainer" containerID="98a126179d6fbc80c0def47519d7a7a3fad4e049faa72c54b859221cf8708f79"
Sep 10 17:43:20 addons-382533 kubelet[2457]: I0910 17:43:20.555398 2457 scope.go:117] "RemoveContainer" containerID="98a126179d6fbc80c0def47519d7a7a3fad4e049faa72c54b859221cf8708f79"
Sep 10 17:43:20 addons-382533 kubelet[2457]: E0910 17:43:20.556066 2457 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 98a126179d6fbc80c0def47519d7a7a3fad4e049faa72c54b859221cf8708f79" containerID="98a126179d6fbc80c0def47519d7a7a3fad4e049faa72c54b859221cf8708f79"
Sep 10 17:43:20 addons-382533 kubelet[2457]: I0910 17:43:20.556112 2457 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"98a126179d6fbc80c0def47519d7a7a3fad4e049faa72c54b859221cf8708f79"} err="failed to get container status \"98a126179d6fbc80c0def47519d7a7a3fad4e049faa72c54b859221cf8708f79\": rpc error: code = Unknown desc = Error response from daemon: No such container: 98a126179d6fbc80c0def47519d7a7a3fad4e049faa72c54b859221cf8708f79"
==> storage-provisioner [858ca178e366] <==
I0910 17:30:46.545569 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0910 17:30:46.560017 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0910 17:30:46.560075 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0910 17:30:46.653846 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0910 17:30:46.654038 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-382533_6f07c7d0-82a4-4229-b1da-65aa5d468078!
I0910 17:30:46.655120 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"91699c70-9f92-4825-b0d0-bd58edff5fcc", APIVersion:"v1", ResourceVersion:"633", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-382533_6f07c7d0-82a4-4229-b1da-65aa5d468078 became leader
I0910 17:30:46.754249 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-382533_6f07c7d0-82a4-4229-b1da-65aa5d468078!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-382533 -n addons-382533
helpers_test.go:261: (dbg) Run: kubectl --context addons-382533 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-382533 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-382533 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-382533/192.168.49.2
Start Time: Tue, 10 Sep 2024 17:34:06 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.28
IPs:
IP: 10.244.0.28
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-xnvsr (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-xnvsr:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m15s default-scheduler Successfully assigned default/busybox to addons-382533
Normal Pulling 7m47s (x4 over 9m14s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m47s (x4 over 9m14s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m47s (x4 over 9m14s) kubelet Error: ErrImagePull
Warning Failed 7m21s (x6 over 9m14s) kubelet Error: ImagePullBackOff
Normal BackOff 4m6s (x20 over 9m14s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (72.46s)