=== RUN TestAddons/parallel/NvidiaDevicePlugin
=== PAUSE TestAddons/parallel/NvidiaDevicePlugin
=== CONT TestAddons/parallel/NvidiaDevicePlugin
addons_test.go:952: (dbg) TestAddons/parallel/NvidiaDevicePlugin: waiting 6m0s for pods matching "name=nvidia-device-plugin-ds" in namespace "kube-system" ...
helpers_test.go:344: "nvidia-device-plugin-daemonset-lmr4m" [3c951f22-d962-4f13-929a-e7a2552f629c] Running
addons_test.go:952: (dbg) TestAddons/parallel/NvidiaDevicePlugin: name=nvidia-device-plugin-ds healthy within 5.004819962s
addons_test.go:955: (dbg) Run: out/minikube-linux-amd64 addons disable nvidia-device-plugin -p addons-214491
addons_test.go:955: (dbg) Non-zero exit: out/minikube-linux-amd64 addons disable nvidia-device-plugin -p addons-214491: exit status 11 (280.175545ms)
-- stdout --
-- /stdout --
** stderr **
X Exiting due to MK_ADDON_DISABLE_PAUSED: disable failed: check paused: list paused: runc: sudo runc --root /run/containerd/runc/k8s.io list -f json: Process exited with status 1
stdout:
stderr:
time="2024-01-31T14:12:28Z" level=error msg="stat /run/containerd/runc/k8s.io/3ff841a74152d483fd0dc29276ab3069cdd544cc0daaab6dfe5c219480f3e9aa: no such file or directory"
*
╭─────────────────────────────────────────────────────────────────────────────────────────────╮
│ │
│ * If the above advice does not help, please let us know: │
│ https://github.com/kubernetes/minikube/issues/new/choose │
│ │
│ * Please run `minikube logs --file=logs.txt` and attach logs.txt to the GitHub issue. │
│ * Please also attach the following file to the GitHub issue: │
│ * - /tmp/minikube_addons_47e1a72799625313bd916979b0f8aa84efd54736_1.log │
│ │
╰─────────────────────────────────────────────────────────────────────────────────────────────╯
** /stderr **
addons_test.go:956: failed to disable nvidia-device-plugin: args "out/minikube-linux-amd64 addons disable nvidia-device-plugin -p addons-214491" : exit status 11
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/NvidiaDevicePlugin]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-214491
helpers_test.go:235: (dbg) docker inspect addons-214491:
-- stdout --
[
{
"Id": "8737208cfcdc52b84ee9a1f2f8218dd282708736b67fe58ea1670ece8d1dd998",
"Created": "2024-01-31T14:10:42.09342074Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 126141,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-01-31T14:10:42.395063379Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:9941de2e064a4a6a7155bfc66cedd2854b8c725b77bb8d4eaf81bef39f951dd7",
"ResolvConfPath": "/var/lib/docker/containers/8737208cfcdc52b84ee9a1f2f8218dd282708736b67fe58ea1670ece8d1dd998/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/8737208cfcdc52b84ee9a1f2f8218dd282708736b67fe58ea1670ece8d1dd998/hostname",
"HostsPath": "/var/lib/docker/containers/8737208cfcdc52b84ee9a1f2f8218dd282708736b67fe58ea1670ece8d1dd998/hosts",
"LogPath": "/var/lib/docker/containers/8737208cfcdc52b84ee9a1f2f8218dd282708736b67fe58ea1670ece8d1dd998/8737208cfcdc52b84ee9a1f2f8218dd282708736b67fe58ea1670ece8d1dd998-json.log",
"Name": "/addons-214491",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-214491:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-214491",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/915311a8fa0a0eaad3ac7c8fe448c837a5037532661e4c140576dd421e171c66-init/diff:/var/lib/docker/overlay2/5f9b5af8b2f6445fb760404f197bfacc3628584467ea8410c1ba7d01af15f15d/diff",
"MergedDir": "/var/lib/docker/overlay2/915311a8fa0a0eaad3ac7c8fe448c837a5037532661e4c140576dd421e171c66/merged",
"UpperDir": "/var/lib/docker/overlay2/915311a8fa0a0eaad3ac7c8fe448c837a5037532661e4c140576dd421e171c66/diff",
"WorkDir": "/var/lib/docker/overlay2/915311a8fa0a0eaad3ac7c8fe448c837a5037532661e4c140576dd421e171c66/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-214491",
"Source": "/var/lib/docker/volumes/addons-214491/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-214491",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"MacAddress": "02:42:c0:a8:31:02",
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-214491",
"name.minikube.sigs.k8s.io": "addons-214491",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "721b634ba24d68a6d339c2b8a26a29fd97199a451768780bf2ea91a0c6f9b46e",
"SandboxKey": "/var/run/docker/netns/721b634ba24d",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-214491": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": [
"8737208cfcdc",
"addons-214491"
],
"MacAddress": "02:42:c0:a8:31:02",
"NetworkID": "d1e6bce6826f9111e3cf41c720f0114a50da870b1b1e79f7de77451d9eef5e81",
"EndpointID": "1fa5c88a06cd2b1b9192851624c3815808587e69d3a913c081a904702f22870a",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DriverOpts": null,
"DNSNames": [
"addons-214491",
"8737208cfcdc"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-214491 -n addons-214491
helpers_test.go:244: <<< TestAddons/parallel/NvidiaDevicePlugin FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/NvidiaDevicePlugin]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-214491 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-amd64 -p addons-214491 logs -n 25: (1.337381453s)
helpers_test.go:252: TestAddons/parallel/NvidiaDevicePlugin logs:
-- stdout --
==> Audit <==
|---------|--------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|--------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| start | -o=json --download-only | download-only-256653 | jenkins | v1.32.0 | 31 Jan 24 14:09 UTC | |
| | -p download-only-256653 | | | | | |
| | --force --alsologtostderr | | | | | |
| | --kubernetes-version=v1.16.0 | | | | | |
| | --container-runtime=containerd | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| delete | --all | minikube | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| delete | -p download-only-256653 | download-only-256653 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| start | -o=json --download-only | download-only-389052 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | |
| | -p download-only-389052 | | | | | |
| | --force --alsologtostderr | | | | | |
| | --kubernetes-version=v1.28.4 | | | | | |
| | --container-runtime=containerd | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| delete | --all | minikube | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| delete | -p download-only-389052 | download-only-389052 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| start | -o=json --download-only | download-only-755607 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | |
| | -p download-only-755607 | | | | | |
| | --force --alsologtostderr | | | | | |
| | --kubernetes-version=v1.29.0-rc.2 | | | | | |
| | --container-runtime=containerd | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| delete | --all | minikube | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| delete | -p download-only-755607 | download-only-755607 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| delete | -p download-only-256653 | download-only-256653 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| delete | -p download-only-389052 | download-only-389052 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| delete | -p download-only-755607 | download-only-755607 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| start | --download-only -p | download-docker-773457 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | |
| | download-docker-773457 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| delete | -p download-docker-773457 | download-docker-773457 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| start | --download-only -p | binary-mirror-953003 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | |
| | binary-mirror-953003 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:43321 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| delete | -p binary-mirror-953003 | binary-mirror-953003 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:10 UTC |
| addons | disable dashboard -p | addons-214491 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | |
| | addons-214491 | | | | | |
| addons | enable dashboard -p | addons-214491 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | |
| | addons-214491 | | | | | |
| start | -p addons-214491 --wait=true | addons-214491 | jenkins | v1.32.0 | 31 Jan 24 14:10 UTC | 31 Jan 24 14:12 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --driver=docker | | | | | |
| | --container-runtime=containerd | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | addons-214491 addons | addons-214491 | jenkins | v1.32.0 | 31 Jan 24 14:12 UTC | 31 Jan 24 14:12 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-214491 addons disable | addons-214491 | jenkins | v1.32.0 | 31 Jan 24 14:12 UTC | 31 Jan 24 14:12 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-214491 | jenkins | v1.32.0 | 31 Jan 24 14:12 UTC | 31 Jan 24 14:12 UTC |
| | addons-214491 | | | | | |
| ip | addons-214491 ip | addons-214491 | jenkins | v1.32.0 | 31 Jan 24 14:12 UTC | 31 Jan 24 14:12 UTC |
| addons | addons-214491 addons disable | addons-214491 | jenkins | v1.32.0 | 31 Jan 24 14:12 UTC | 31 Jan 24 14:12 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-214491 | jenkins | v1.32.0 | 31 Jan 24 14:12 UTC | |
| | -p addons-214491 | | | | | |
|---------|--------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/01/31 14:10:20
Running on machine: ubuntu-20-agent-15
Binary: Built with gc go1.21.6 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0131 14:10:20.239124 125487 out.go:296] Setting OutFile to fd 1 ...
I0131 14:10:20.239334 125487 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0131 14:10:20.239350 125487 out.go:309] Setting ErrFile to fd 2...
I0131 14:10:20.239363 125487 out.go:343] TERM=,COLORTERM=, which probably does not support color
I0131 14:10:20.239629 125487 root.go:338] Updating PATH: /home/jenkins/minikube-integration/18007-117277/.minikube/bin
I0131 14:10:20.240462 125487 out.go:303] Setting JSON to false
I0131 14:10:20.241579 125487 start.go:128] hostinfo: {"hostname":"ubuntu-20-agent-15","uptime":67972,"bootTime":1706642248,"procs":177,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1049-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0131 14:10:20.241664 125487 start.go:138] virtualization: kvm guest
I0131 14:10:20.243921 125487 out.go:177] * [addons-214491] minikube v1.32.0 on Ubuntu 20.04 (kvm/amd64)
I0131 14:10:20.245617 125487 out.go:177] - MINIKUBE_LOCATION=18007
I0131 14:10:20.245679 125487 notify.go:220] Checking for updates...
I0131 14:10:20.246863 125487 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0131 14:10:20.248720 125487 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/18007-117277/kubeconfig
I0131 14:10:20.250273 125487 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/18007-117277/.minikube
I0131 14:10:20.251702 125487 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0131 14:10:20.253040 125487 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0131 14:10:20.254654 125487 driver.go:392] Setting default libvirt URI to qemu:///system
I0131 14:10:20.277838 125487 docker.go:122] docker version: linux-25.0.1:Docker Engine - Community
I0131 14:10:20.277991 125487 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0131 14:10:20.328270 125487 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:31 OomKillDisable:true NGoroutines:51 SystemTime:2024-01-31 14:10:20.31874443 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1049-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86_
64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33648054272 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-15 Labels:[] ExperimentalBuild:false ServerVersion:25.0.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:a1496014c916f9e62104b33d1bb5bd03b0858e59 Expected:a1496014c916f9e62104b33d1bb5bd03b0858e59} RuncCommit:{ID:v1.1.11-0-g4bccb38 Expected:v1.1.11-0-g4bccb38} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErr
ors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.12.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.24.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0131 14:10:20.328474 125487 docker.go:295] overlay module found
I0131 14:10:20.330457 125487 out.go:177] * Using the docker driver based on user configuration
I0131 14:10:20.331840 125487 start.go:298] selected driver: docker
I0131 14:10:20.331854 125487 start.go:902] validating driver "docker" against <nil>
I0131 14:10:20.331867 125487 start.go:913] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0131 14:10:20.332730 125487 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0131 14:10:20.386750 125487 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:31 OomKillDisable:true NGoroutines:51 SystemTime:2024-01-31 14:10:20.376984302 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1049-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33648054272 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-15 Labels:[] ExperimentalBuild:false ServerVersion:25.0.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:a1496014c916f9e62104b33d1bb5bd03b0858e59 Expected:a1496014c916f9e62104b33d1bb5bd03b0858e59} RuncCommit:{ID:v1.1.11-0-g4bccb38 Expected:v1.1.11-0-g4bccb38} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.12.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.24.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0131 14:10:20.386913 125487 start_flags.go:307] no existing cluster config was found, will generate one from the flags
I0131 14:10:20.387129 125487 start_flags.go:927] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0131 14:10:20.388762 125487 out.go:177] * Using Docker driver with root privileges
I0131 14:10:20.390236 125487 cni.go:84] Creating CNI manager for ""
I0131 14:10:20.390265 125487 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0131 14:10:20.390280 125487 start_flags.go:316] Found "CNI" CNI - setting NetworkPlugin=cni
I0131 14:10:20.390302 125487 start_flags.go:321] config:
{Name:addons-214491 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.4 ClusterName:addons-214491 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containe
rd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs:}
I0131 14:10:20.391806 125487 out.go:177] * Starting control plane node addons-214491 in cluster addons-214491
I0131 14:10:20.392956 125487 cache.go:121] Beginning downloading kic base image for docker with containerd
I0131 14:10:20.394289 125487 out.go:177] * Pulling base image v0.0.42-1704759386-17866 ...
I0131 14:10:20.395440 125487 preload.go:132] Checking if preload exists for k8s version v1.28.4 and runtime containerd
I0131 14:10:20.395492 125487 preload.go:148] Found local preload: /home/jenkins/minikube-integration/18007-117277/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.4-containerd-overlay2-amd64.tar.lz4
I0131 14:10:20.395507 125487 cache.go:56] Caching tarball of preloaded images
I0131 14:10:20.395544 125487 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 in local docker daemon
I0131 14:10:20.395611 125487 preload.go:174] Found /home/jenkins/minikube-integration/18007-117277/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.4-containerd-overlay2-amd64.tar.lz4 in cache, skipping download
I0131 14:10:20.395629 125487 cache.go:59] Finished verifying existence of preloaded tar for v1.28.4 on containerd
I0131 14:10:20.396021 125487 profile.go:148] Saving config to /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/config.json ...
I0131 14:10:20.396060 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/config.json: {Name:mk52e93e7af9c1113f1244badd52ecd8e57ddd60 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:20.412921 125487 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 to local cache
I0131 14:10:20.413141 125487 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 in local cache directory
I0131 14:10:20.413170 125487 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 in local cache directory, skipping pull
I0131 14:10:20.413178 125487 image.go:105] gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 exists in cache, skipping pull
I0131 14:10:20.413187 125487 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 as a tarball
I0131 14:10:20.413194 125487 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 from local cache
I0131 14:10:33.256880 125487 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 from cached tarball
I0131 14:10:33.256945 125487 cache.go:194] Successfully downloaded all kic artifacts
I0131 14:10:33.256998 125487 start.go:365] acquiring machines lock for addons-214491: {Name:mkbc57ae4120ed19530588052ad7c3467a369857 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0131 14:10:33.257127 125487 start.go:369] acquired machines lock for "addons-214491" in 100.261µs
I0131 14:10:33.257155 125487 start.go:93] Provisioning new machine with config: &{Name:addons-214491 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.4 ClusterName:addons-214491 Namespace:default APIServerName:minikubeCA A
PIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.28.4 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:
false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs:} &{Name: IP: Port:8443 KubernetesVersion:v1.28.4 ContainerRuntime:containerd ControlPlane:true Worker:true}
I0131 14:10:33.257281 125487 start.go:125] createHost starting for "" (driver="docker")
I0131 14:10:33.340725 125487 out.go:204] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0131 14:10:33.341107 125487 start.go:159] libmachine.API.Create for "addons-214491" (driver="docker")
I0131 14:10:33.341148 125487 client.go:168] LocalClient.Create starting
I0131 14:10:33.341326 125487 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/18007-117277/.minikube/certs/ca.pem
I0131 14:10:33.607224 125487 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/18007-117277/.minikube/certs/cert.pem
I0131 14:10:33.702617 125487 cli_runner.go:164] Run: docker network inspect addons-214491 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0131 14:10:33.720095 125487 cli_runner.go:211] docker network inspect addons-214491 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0131 14:10:33.720203 125487 network_create.go:281] running [docker network inspect addons-214491] to gather additional debugging logs...
I0131 14:10:33.720228 125487 cli_runner.go:164] Run: docker network inspect addons-214491
W0131 14:10:33.737516 125487 cli_runner.go:211] docker network inspect addons-214491 returned with exit code 1
I0131 14:10:33.737554 125487 network_create.go:284] error running [docker network inspect addons-214491]: docker network inspect addons-214491: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-214491 not found
I0131 14:10:33.737577 125487 network_create.go:286] output of [docker network inspect addons-214491]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-214491 not found
** /stderr **
I0131 14:10:33.737695 125487 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0131 14:10:33.757772 125487 network.go:209] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc002855640}
I0131 14:10:33.757851 125487 network_create.go:124] attempt to create docker network addons-214491 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0131 14:10:33.757935 125487 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-214491 addons-214491
I0131 14:10:34.006693 125487 network_create.go:108] docker network addons-214491 192.168.49.0/24 created
I0131 14:10:34.006726 125487 kic.go:121] calculated static IP "192.168.49.2" for the "addons-214491" container
I0131 14:10:34.006798 125487 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0131 14:10:34.022653 125487 cli_runner.go:164] Run: docker volume create addons-214491 --label name.minikube.sigs.k8s.io=addons-214491 --label created_by.minikube.sigs.k8s.io=true
I0131 14:10:34.127442 125487 oci.go:103] Successfully created a docker volume addons-214491
I0131 14:10:34.127561 125487 cli_runner.go:164] Run: docker run --rm --name addons-214491-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-214491 --entrypoint /usr/bin/test -v addons-214491:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 -d /var/lib
I0131 14:10:36.520531 125487 cli_runner.go:217] Completed: docker run --rm --name addons-214491-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-214491 --entrypoint /usr/bin/test -v addons-214491:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 -d /var/lib: (2.392920929s)
I0131 14:10:36.520566 125487 oci.go:107] Successfully prepared a docker volume addons-214491
I0131 14:10:36.520586 125487 preload.go:132] Checking if preload exists for k8s version v1.28.4 and runtime containerd
I0131 14:10:36.520609 125487 kic.go:194] Starting extracting preloaded images to volume ...
I0131 14:10:36.520669 125487 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/18007-117277/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.4-containerd-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-214491:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 -I lz4 -xf /preloaded.tar -C /extractDir
I0131 14:10:42.023161 125487 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/18007-117277/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.28.4-containerd-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-214491:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 -I lz4 -xf /preloaded.tar -C /extractDir: (5.502437986s)
I0131 14:10:42.023200 125487 kic.go:203] duration metric: took 5.502588 seconds to extract preloaded images to volume
W0131 14:10:42.023380 125487 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0131 14:10:42.023493 125487 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0131 14:10:42.078520 125487 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-214491 --name addons-214491 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-214491 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-214491 --network addons-214491 --ip 192.168.49.2 --volume addons-214491:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0
I0131 14:10:42.403293 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Running}}
I0131 14:10:42.422808 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:10:42.443613 125487 cli_runner.go:164] Run: docker exec addons-214491 stat /var/lib/dpkg/alternatives/iptables
I0131 14:10:42.489084 125487 oci.go:144] the created container "addons-214491" has a running status.
I0131 14:10:42.489116 125487 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa...
I0131 14:10:42.673543 125487 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0131 14:10:42.693980 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:10:42.714166 125487 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0131 14:10:42.714201 125487 kic_runner.go:114] Args: [docker exec --privileged addons-214491 chown docker:docker /home/docker/.ssh/authorized_keys]
I0131 14:10:42.769127 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:10:42.787116 125487 machine.go:88] provisioning docker machine ...
I0131 14:10:42.787167 125487 ubuntu.go:169] provisioning hostname "addons-214491"
I0131 14:10:42.787238 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:10:42.808226 125487 main.go:141] libmachine: Using SSH client type: native
I0131 14:10:42.808890 125487 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x80aa00] 0x80d6e0 <nil> [] 0s} 127.0.0.1 32772 <nil> <nil>}
I0131 14:10:42.808924 125487 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-214491 && echo "addons-214491" | sudo tee /etc/hostname
I0131 14:10:42.809851 125487 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: read tcp 127.0.0.1:55380->127.0.0.1:32772: read: connection reset by peer
I0131 14:10:45.959186 125487 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-214491
I0131 14:10:45.959304 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:10:45.977243 125487 main.go:141] libmachine: Using SSH client type: native
I0131 14:10:45.977653 125487 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x80aa00] 0x80d6e0 <nil> [] 0s} 127.0.0.1 32772 <nil> <nil>}
I0131 14:10:45.977673 125487 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-214491' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-214491/g' /etc/hosts;
else
echo '127.0.1.1 addons-214491' | sudo tee -a /etc/hosts;
fi
fi
I0131 14:10:46.110653 125487 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0131 14:10:46.110697 125487 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/18007-117277/.minikube CaCertPath:/home/jenkins/minikube-integration/18007-117277/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/18007-117277/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/18007-117277/.minikube}
I0131 14:10:46.110732 125487 ubuntu.go:177] setting up certificates
I0131 14:10:46.110751 125487 provision.go:83] configureAuth start
I0131 14:10:46.110847 125487 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-214491
I0131 14:10:46.130659 125487 provision.go:138] copyHostCerts
I0131 14:10:46.130759 125487 exec_runner.go:151] cp: /home/jenkins/minikube-integration/18007-117277/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/18007-117277/.minikube/ca.pem (1078 bytes)
I0131 14:10:46.130910 125487 exec_runner.go:151] cp: /home/jenkins/minikube-integration/18007-117277/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/18007-117277/.minikube/cert.pem (1123 bytes)
I0131 14:10:46.130984 125487 exec_runner.go:151] cp: /home/jenkins/minikube-integration/18007-117277/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/18007-117277/.minikube/key.pem (1679 bytes)
I0131 14:10:46.131056 125487 provision.go:112] generating server cert: /home/jenkins/minikube-integration/18007-117277/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/18007-117277/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/18007-117277/.minikube/certs/ca-key.pem org=jenkins.addons-214491 san=[192.168.49.2 127.0.0.1 localhost 127.0.0.1 minikube addons-214491]
I0131 14:10:46.252263 125487 provision.go:172] copyRemoteCerts
I0131 14:10:46.252335 125487 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0131 14:10:46.252377 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:10:46.270276 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:10:46.366275 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0131 14:10:46.389093 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/machines/server.pem --> /etc/docker/server.pem (1216 bytes)
I0131 14:10:46.411185 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0131 14:10:46.435375 125487 provision.go:86] duration metric: configureAuth took 324.60332ms
I0131 14:10:46.435404 125487 ubuntu.go:193] setting minikube options for container-runtime
I0131 14:10:46.435580 125487 config.go:182] Loaded profile config "addons-214491": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.4
I0131 14:10:46.435592 125487 machine.go:91] provisioned docker machine in 3.648452061s
I0131 14:10:46.435599 125487 client.go:171] LocalClient.Create took 13.094444828s
I0131 14:10:46.435617 125487 start.go:167] duration metric: libmachine.API.Create for "addons-214491" took 13.09451664s
I0131 14:10:46.435626 125487 start.go:300] post-start starting for "addons-214491" (driver="docker")
I0131 14:10:46.435639 125487 start.go:329] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0131 14:10:46.435681 125487 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0131 14:10:46.435722 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:10:46.454278 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:10:46.551117 125487 ssh_runner.go:195] Run: cat /etc/os-release
I0131 14:10:46.554846 125487 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0131 14:10:46.554879 125487 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0131 14:10:46.554887 125487 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0131 14:10:46.554895 125487 info.go:137] Remote host: Ubuntu 22.04.3 LTS
I0131 14:10:46.554908 125487 filesync.go:126] Scanning /home/jenkins/minikube-integration/18007-117277/.minikube/addons for local assets ...
I0131 14:10:46.554995 125487 filesync.go:126] Scanning /home/jenkins/minikube-integration/18007-117277/.minikube/files for local assets ...
I0131 14:10:46.555022 125487 start.go:303] post-start completed in 119.389166ms
I0131 14:10:46.555317 125487 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-214491
I0131 14:10:46.572599 125487 profile.go:148] Saving config to /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/config.json ...
I0131 14:10:46.572874 125487 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0131 14:10:46.572916 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:10:46.589150 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:10:46.683020 125487 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0131 14:10:46.688115 125487 start.go:128] duration metric: createHost completed in 13.430804476s
I0131 14:10:46.688156 125487 start.go:83] releasing machines lock for "addons-214491", held for 13.431016468s
I0131 14:10:46.688245 125487 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-214491
I0131 14:10:46.707620 125487 ssh_runner.go:195] Run: cat /version.json
I0131 14:10:46.707701 125487 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0131 14:10:46.707719 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:10:46.707810 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:10:46.726478 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:10:46.726797 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:10:46.911118 125487 ssh_runner.go:195] Run: systemctl --version
I0131 14:10:46.915630 125487 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0131 14:10:46.920121 125487 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0131 14:10:46.943786 125487 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0131 14:10:46.943860 125487 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%!p(MISSING), " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0131 14:10:46.968723 125487 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0131 14:10:46.968756 125487 start.go:475] detecting cgroup driver to use...
I0131 14:10:46.968792 125487 detect.go:196] detected "cgroupfs" cgroup driver on host os
I0131 14:10:46.968837 125487 ssh_runner.go:195] Run: sudo systemctl stop -f crio
I0131 14:10:46.979968 125487 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0131 14:10:46.989769 125487 docker.go:217] disabling cri-docker service (if available) ...
I0131 14:10:46.989815 125487 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.socket
I0131 14:10:47.002041 125487 ssh_runner.go:195] Run: sudo systemctl stop -f cri-docker.service
I0131 14:10:47.014440 125487 ssh_runner.go:195] Run: sudo systemctl disable cri-docker.socket
I0131 14:10:47.086449 125487 ssh_runner.go:195] Run: sudo systemctl mask cri-docker.service
I0131 14:10:47.162776 125487 docker.go:233] disabling docker service ...
I0131 14:10:47.162841 125487 ssh_runner.go:195] Run: sudo systemctl stop -f docker.socket
I0131 14:10:47.181593 125487 ssh_runner.go:195] Run: sudo systemctl stop -f docker.service
I0131 14:10:47.191843 125487 ssh_runner.go:195] Run: sudo systemctl disable docker.socket
I0131 14:10:47.270227 125487 ssh_runner.go:195] Run: sudo systemctl mask docker.service
I0131 14:10:47.345899 125487 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service docker
I0131 14:10:47.356329 125487 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0131 14:10:47.370883 125487 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.9"|' /etc/containerd/config.toml"
I0131 14:10:47.379738 125487 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0131 14:10:47.388969 125487 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0131 14:10:47.389037 125487 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0131 14:10:47.397985 125487 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0131 14:10:47.406571 125487 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0131 14:10:47.415295 125487 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0131 14:10:47.423763 125487 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0131 14:10:47.431671 125487 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0131 14:10:47.440198 125487 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0131 14:10:47.447942 125487 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0131 14:10:47.455431 125487 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0131 14:10:47.529201 125487 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0131 14:10:47.628443 125487 start.go:522] Will wait 60s for socket path /run/containerd/containerd.sock
I0131 14:10:47.628536 125487 ssh_runner.go:195] Run: stat /run/containerd/containerd.sock
I0131 14:10:47.632099 125487 start.go:543] Will wait 60s for crictl version
I0131 14:10:47.632157 125487 ssh_runner.go:195] Run: which crictl
I0131 14:10:47.635295 125487 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0131 14:10:47.672212 125487 start.go:559] Version: 0.1.0
RuntimeName: containerd
RuntimeVersion: 1.6.26
RuntimeApiVersion: v1
I0131 14:10:47.672303 125487 ssh_runner.go:195] Run: containerd --version
I0131 14:10:47.699694 125487 ssh_runner.go:195] Run: containerd --version
I0131 14:10:47.728950 125487 out.go:177] * Preparing Kubernetes v1.28.4 on containerd 1.6.26 ...
I0131 14:10:47.730225 125487 cli_runner.go:164] Run: docker network inspect addons-214491 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0131 14:10:47.748523 125487 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0131 14:10:47.752873 125487 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0131 14:10:47.765523 125487 preload.go:132] Checking if preload exists for k8s version v1.28.4 and runtime containerd
I0131 14:10:47.765637 125487 ssh_runner.go:195] Run: sudo crictl images --output json
I0131 14:10:47.801882 125487 containerd.go:612] all images are preloaded for containerd runtime.
I0131 14:10:47.801904 125487 containerd.go:519] Images already preloaded, skipping extraction
I0131 14:10:47.801952 125487 ssh_runner.go:195] Run: sudo crictl images --output json
I0131 14:10:47.835469 125487 containerd.go:612] all images are preloaded for containerd runtime.
I0131 14:10:47.835494 125487 cache_images.go:84] Images are preloaded, skipping loading
I0131 14:10:47.835546 125487 ssh_runner.go:195] Run: sudo crictl info
I0131 14:10:47.870282 125487 cni.go:84] Creating CNI manager for ""
I0131 14:10:47.870307 125487 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0131 14:10:47.870327 125487 kubeadm.go:87] Using pod CIDR: 10.244.0.0/16
I0131 14:10:47.870345 125487 kubeadm.go:176] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.28.4 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-214491 NodeName:addons-214491 DNSDomain:cluster.local CRISocket:/run/containerd/containerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc
/kubernetes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0131 14:10:47.870468 125487 kubeadm.go:181] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///run/containerd/containerd.sock
name: "addons-214491"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.28.4
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%!"(MISSING)
nodefs.inodesFree: "0%!"(MISSING)
imagefs.available: "0%!"(MISSING)
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0131 14:10:47.870530 125487 kubeadm.go:976] kubelet [Unit]
Wants=containerd.service
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.28.4/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --container-runtime-endpoint=unix:///run/containerd/containerd.sock --hostname-override=addons-214491 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.28.4 ClusterName:addons-214491 Namespace:default APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:}
I0131 14:10:47.870584 125487 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.28.4
I0131 14:10:47.879401 125487 binaries.go:44] Found k8s binaries, skipping transfer
I0131 14:10:47.879475 125487 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0131 14:10:47.888156 125487 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (385 bytes)
I0131 14:10:47.907027 125487 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0131 14:10:47.925683 125487 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2102 bytes)
I0131 14:10:47.945625 125487 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0131 14:10:47.949820 125487 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0131 14:10:47.962235 125487 certs.go:56] Setting up /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491 for IP: 192.168.49.2
I0131 14:10:47.962281 125487 certs.go:190] acquiring lock for shared ca certs: {Name:mkeb54e6602fe1d63447effea679de9e9af3fb32 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:47.962436 125487 certs.go:204] generating minikubeCA CA: /home/jenkins/minikube-integration/18007-117277/.minikube/ca.key
I0131 14:10:48.028098 125487 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/18007-117277/.minikube/ca.crt ...
I0131 14:10:48.028127 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/ca.crt: {Name:mk6c1f82a770c32d1a5ebed3b7d62473053df569 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:48.028284 125487 crypto.go:164] Writing key to /home/jenkins/minikube-integration/18007-117277/.minikube/ca.key ...
I0131 14:10:48.028307 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/ca.key: {Name:mk7cdc03b4b70e2d67861f62263e1d14dfacc982 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:48.028389 125487 certs.go:204] generating proxyClientCA CA: /home/jenkins/minikube-integration/18007-117277/.minikube/proxy-client-ca.key
I0131 14:10:48.096211 125487 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/18007-117277/.minikube/proxy-client-ca.crt ...
I0131 14:10:48.096246 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/proxy-client-ca.crt: {Name:mkc6e03e598486fc42da1a3749132bb565940a3c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:48.096430 125487 crypto.go:164] Writing key to /home/jenkins/minikube-integration/18007-117277/.minikube/proxy-client-ca.key ...
I0131 14:10:48.096442 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/proxy-client-ca.key: {Name:mk390459cac59acd892564bf4f2438e7089b8ba7 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:48.096551 125487 certs.go:319] generating minikube-user signed cert: /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/client.key
I0131 14:10:48.096564 125487 crypto.go:68] Generating cert /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/client.crt with IP's: []
I0131 14:10:48.176090 125487 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/client.crt ...
I0131 14:10:48.176126 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/client.crt: {Name:mkaf1bc6cad3f24c99fb866d0261c634911a88ad Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:48.176291 125487 crypto.go:164] Writing key to /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/client.key ...
I0131 14:10:48.176302 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/client.key: {Name:mkdc3b2e0e2e45e78eb4f47bba9364a3e0e1adfc Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:48.176383 125487 certs.go:319] generating minikube signed cert: /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.key.dd3b5fb2
I0131 14:10:48.176401 125487 crypto.go:68] Generating cert /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.crt.dd3b5fb2 with IP's: [192.168.49.2 10.96.0.1 127.0.0.1 10.0.0.1]
I0131 14:10:48.297087 125487 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.crt.dd3b5fb2 ...
I0131 14:10:48.297133 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.crt.dd3b5fb2: {Name:mk880512e4de89ef95e01190d3858772c87d57f9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:48.297356 125487 crypto.go:164] Writing key to /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.key.dd3b5fb2 ...
I0131 14:10:48.297378 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.key.dd3b5fb2: {Name:mka2d2f7204800413533ebfaf9388e213453580c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:48.297543 125487 certs.go:337] copying /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.crt.dd3b5fb2 -> /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.crt
I0131 14:10:48.297654 125487 certs.go:341] copying /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.key.dd3b5fb2 -> /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.key
I0131 14:10:48.297722 125487 certs.go:319] generating aggregator signed cert: /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/proxy-client.key
I0131 14:10:48.297747 125487 crypto.go:68] Generating cert /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/proxy-client.crt with IP's: []
I0131 14:10:48.431932 125487 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/proxy-client.crt ...
I0131 14:10:48.431968 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/proxy-client.crt: {Name:mkf20743cd84e3ec108055df042f5d875327c93a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:48.432167 125487 crypto.go:164] Writing key to /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/proxy-client.key ...
I0131 14:10:48.432189 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/proxy-client.key: {Name:mk73fdc86d07f0e5cd2e26bc026e4cabff8c4358 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:10:48.432450 125487 certs.go:437] found cert: /home/jenkins/minikube-integration/18007-117277/.minikube/certs/home/jenkins/minikube-integration/18007-117277/.minikube/certs/ca-key.pem (1679 bytes)
I0131 14:10:48.432507 125487 certs.go:437] found cert: /home/jenkins/minikube-integration/18007-117277/.minikube/certs/home/jenkins/minikube-integration/18007-117277/.minikube/certs/ca.pem (1078 bytes)
I0131 14:10:48.432549 125487 certs.go:437] found cert: /home/jenkins/minikube-integration/18007-117277/.minikube/certs/home/jenkins/minikube-integration/18007-117277/.minikube/certs/cert.pem (1123 bytes)
I0131 14:10:48.432584 125487 certs.go:437] found cert: /home/jenkins/minikube-integration/18007-117277/.minikube/certs/home/jenkins/minikube-integration/18007-117277/.minikube/certs/key.pem (1679 bytes)
I0131 14:10:48.433194 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1399 bytes)
I0131 14:10:48.456104 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0131 14:10:48.477232 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0131 14:10:48.498530 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/profiles/addons-214491/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I0131 14:10:48.519488 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0131 14:10:48.540612 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0131 14:10:48.561983 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0131 14:10:48.582983 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0131 14:10:48.603865 125487 ssh_runner.go:362] scp /home/jenkins/minikube-integration/18007-117277/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0131 14:10:48.625420 125487 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0131 14:10:48.641693 125487 ssh_runner.go:195] Run: openssl version
I0131 14:10:48.646817 125487 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0131 14:10:48.655219 125487 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0131 14:10:48.658268 125487 certs.go:480] hashing: -rw-r--r-- 1 root root 1111 Jan 31 14:10 /usr/share/ca-certificates/minikubeCA.pem
I0131 14:10:48.658323 125487 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0131 14:10:48.664451 125487 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0131 14:10:48.672791 125487 ssh_runner.go:195] Run: ls /var/lib/minikube/certs/etcd
I0131 14:10:48.675789 125487 certs.go:353] certs directory doesn't exist, likely first start: ls /var/lib/minikube/certs/etcd: Process exited with status 2
stdout:
stderr:
ls: cannot access '/var/lib/minikube/certs/etcd': No such file or directory
I0131 14:10:48.675841 125487 kubeadm.go:404] StartCluster: {Name:addons-214491 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.42-1704759386-17866@sha256:8c3c33047f9bc285e1f5f2a5aa14744a2fe04c58478f02f77b06169dea8dd3f0 Memory:4000 CPUs:2 DiskSize:20000 VMDriver: Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:0 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.28.4 ClusterName:addons-214491 Namespace:default APIServerName:minikubeCA APIServerNames:[] APISe
rverIPs:[] DNSDomain:cluster.local ContainerRuntime:containerd CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI: NodeIP: NodePort:8443 NodeName:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.28.4 ContainerRuntime:containerd ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false Disa
bleMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs:}
I0131 14:10:48.675923 125487 cri.go:54] listing CRI containers in root /run/containerd/runc/k8s.io: {State:paused Name: Namespaces:[kube-system]}
I0131 14:10:48.675962 125487 ssh_runner.go:195] Run: sudo -s eval "crictl ps -a --quiet --label io.kubernetes.pod.namespace=kube-system"
I0131 14:10:48.708938 125487 cri.go:89] found id: ""
I0131 14:10:48.709004 125487 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0131 14:10:48.717428 125487 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0131 14:10:48.725359 125487 kubeadm.go:226] ignoring SystemVerification for kubeadm because of docker driver
I0131 14:10:48.725418 125487 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0131 14:10:48.733063 125487 kubeadm.go:152] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0131 14:10:48.733134 125487 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.28.4:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0131 14:10:48.777228 125487 kubeadm.go:322] [init] Using Kubernetes version: v1.28.4
I0131 14:10:48.777322 125487 kubeadm.go:322] [preflight] Running pre-flight checks
I0131 14:10:48.811965 125487 kubeadm.go:322] [preflight] The system verification failed. Printing the output from the verification:
I0131 14:10:48.812030 125487 kubeadm.go:322] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1049-gcp[0m
I0131 14:10:48.812061 125487 kubeadm.go:322] [0;37mOS[0m: [0;32mLinux[0m
I0131 14:10:48.812105 125487 kubeadm.go:322] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0131 14:10:48.812159 125487 kubeadm.go:322] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0131 14:10:48.812212 125487 kubeadm.go:322] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0131 14:10:48.812251 125487 kubeadm.go:322] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0131 14:10:48.812296 125487 kubeadm.go:322] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0131 14:10:48.812343 125487 kubeadm.go:322] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0131 14:10:48.812381 125487 kubeadm.go:322] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0131 14:10:48.812421 125487 kubeadm.go:322] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0131 14:10:48.812477 125487 kubeadm.go:322] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0131 14:10:48.874151 125487 kubeadm.go:322] [preflight] Pulling images required for setting up a Kubernetes cluster
I0131 14:10:48.874277 125487 kubeadm.go:322] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0131 14:10:48.874372 125487 kubeadm.go:322] [preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
I0131 14:10:49.065737 125487 kubeadm.go:322] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0131 14:10:49.068481 125487 out.go:204] - Generating certificates and keys ...
I0131 14:10:49.068587 125487 kubeadm.go:322] [certs] Using existing ca certificate authority
I0131 14:10:49.068678 125487 kubeadm.go:322] [certs] Using existing apiserver certificate and key on disk
I0131 14:10:49.143900 125487 kubeadm.go:322] [certs] Generating "apiserver-kubelet-client" certificate and key
I0131 14:10:49.229617 125487 kubeadm.go:322] [certs] Generating "front-proxy-ca" certificate and key
I0131 14:10:49.503876 125487 kubeadm.go:322] [certs] Generating "front-proxy-client" certificate and key
I0131 14:10:49.593614 125487 kubeadm.go:322] [certs] Generating "etcd/ca" certificate and key
I0131 14:10:49.841803 125487 kubeadm.go:322] [certs] Generating "etcd/server" certificate and key
I0131 14:10:49.841998 125487 kubeadm.go:322] [certs] etcd/server serving cert is signed for DNS names [addons-214491 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0131 14:10:49.889192 125487 kubeadm.go:322] [certs] Generating "etcd/peer" certificate and key
I0131 14:10:49.889369 125487 kubeadm.go:322] [certs] etcd/peer serving cert is signed for DNS names [addons-214491 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0131 14:10:50.099045 125487 kubeadm.go:322] [certs] Generating "etcd/healthcheck-client" certificate and key
I0131 14:10:50.202427 125487 kubeadm.go:322] [certs] Generating "apiserver-etcd-client" certificate and key
I0131 14:10:50.303317 125487 kubeadm.go:322] [certs] Generating "sa" key and public key
I0131 14:10:50.303391 125487 kubeadm.go:322] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0131 14:10:50.480627 125487 kubeadm.go:322] [kubeconfig] Writing "admin.conf" kubeconfig file
I0131 14:10:50.573920 125487 kubeadm.go:322] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0131 14:10:50.781732 125487 kubeadm.go:322] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0131 14:10:50.916184 125487 kubeadm.go:322] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0131 14:10:50.917629 125487 kubeadm.go:322] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0131 14:10:50.919883 125487 kubeadm.go:322] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0131 14:10:50.922032 125487 out.go:204] - Booting up control plane ...
I0131 14:10:50.922173 125487 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0131 14:10:50.922295 125487 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0131 14:10:50.922385 125487 kubeadm.go:322] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0131 14:10:50.934072 125487 kubeadm.go:322] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0131 14:10:50.934919 125487 kubeadm.go:322] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0131 14:10:50.934991 125487 kubeadm.go:322] [kubelet-start] Starting the kubelet
I0131 14:10:51.015347 125487 kubeadm.go:322] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests". This can take up to 4m0s
I0131 14:10:56.017840 125487 kubeadm.go:322] [apiclient] All control plane components are healthy after 5.002597 seconds
I0131 14:10:56.017949 125487 kubeadm.go:322] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0131 14:10:56.031951 125487 kubeadm.go:322] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0131 14:10:56.552919 125487 kubeadm.go:322] [upload-certs] Skipping phase. Please see --upload-certs
I0131 14:10:56.553119 125487 kubeadm.go:322] [mark-control-plane] Marking the node addons-214491 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0131 14:10:57.063391 125487 kubeadm.go:322] [bootstrap-token] Using token: b06pqs.qtbob5qs89luanua
I0131 14:10:57.064635 125487 out.go:204] - Configuring RBAC rules ...
I0131 14:10:57.064791 125487 kubeadm.go:322] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0131 14:10:57.070016 125487 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0131 14:10:57.078073 125487 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0131 14:10:57.081596 125487 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0131 14:10:57.086376 125487 kubeadm.go:322] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0131 14:10:57.089775 125487 kubeadm.go:322] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0131 14:10:57.101384 125487 kubeadm.go:322] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0131 14:10:57.402550 125487 kubeadm.go:322] [addons] Applied essential addon: CoreDNS
I0131 14:10:57.509253 125487 kubeadm.go:322] [addons] Applied essential addon: kube-proxy
I0131 14:10:57.510459 125487 kubeadm.go:322]
I0131 14:10:57.510551 125487 kubeadm.go:322] Your Kubernetes control-plane has initialized successfully!
I0131 14:10:57.510561 125487 kubeadm.go:322]
I0131 14:10:57.510686 125487 kubeadm.go:322] To start using your cluster, you need to run the following as a regular user:
I0131 14:10:57.510709 125487 kubeadm.go:322]
I0131 14:10:57.510737 125487 kubeadm.go:322] mkdir -p $HOME/.kube
I0131 14:10:57.510810 125487 kubeadm.go:322] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0131 14:10:57.510871 125487 kubeadm.go:322] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0131 14:10:57.510880 125487 kubeadm.go:322]
I0131 14:10:57.510951 125487 kubeadm.go:322] Alternatively, if you are the root user, you can run:
I0131 14:10:57.510958 125487 kubeadm.go:322]
I0131 14:10:57.511023 125487 kubeadm.go:322] export KUBECONFIG=/etc/kubernetes/admin.conf
I0131 14:10:57.511030 125487 kubeadm.go:322]
I0131 14:10:57.511093 125487 kubeadm.go:322] You should now deploy a pod network to the cluster.
I0131 14:10:57.511188 125487 kubeadm.go:322] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0131 14:10:57.511280 125487 kubeadm.go:322] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0131 14:10:57.511288 125487 kubeadm.go:322]
I0131 14:10:57.511399 125487 kubeadm.go:322] You can now join any number of control-plane nodes by copying certificate authorities
I0131 14:10:57.511496 125487 kubeadm.go:322] and service account keys on each node and then running the following as root:
I0131 14:10:57.511503 125487 kubeadm.go:322]
I0131 14:10:57.511598 125487 kubeadm.go:322] kubeadm join control-plane.minikube.internal:8443 --token b06pqs.qtbob5qs89luanua \
I0131 14:10:57.511741 125487 kubeadm.go:322] --discovery-token-ca-cert-hash sha256:919e83d6c7eedee9017f84f7359a65c9a85e797008d7d7c0cf490ef8044657cb \
I0131 14:10:57.511766 125487 kubeadm.go:322] --control-plane
I0131 14:10:57.511773 125487 kubeadm.go:322]
I0131 14:10:57.511881 125487 kubeadm.go:322] Then you can join any number of worker nodes by running the following on each as root:
I0131 14:10:57.511888 125487 kubeadm.go:322]
I0131 14:10:57.512020 125487 kubeadm.go:322] kubeadm join control-plane.minikube.internal:8443 --token b06pqs.qtbob5qs89luanua \
I0131 14:10:57.512192 125487 kubeadm.go:322] --discovery-token-ca-cert-hash sha256:919e83d6c7eedee9017f84f7359a65c9a85e797008d7d7c0cf490ef8044657cb
I0131 14:10:57.515022 125487 kubeadm.go:322] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1049-gcp\n", err: exit status 1
I0131 14:10:57.515176 125487 kubeadm.go:322] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0131 14:10:57.515212 125487 cni.go:84] Creating CNI manager for ""
I0131 14:10:57.515231 125487 cni.go:143] "docker" driver + "containerd" runtime found, recommending kindnet
I0131 14:10:57.516994 125487 out.go:177] * Configuring CNI (Container Networking Interface) ...
I0131 14:10:57.518350 125487 ssh_runner.go:195] Run: stat /opt/cni/bin/portmap
I0131 14:10:57.524188 125487 cni.go:182] applying CNI manifest using /var/lib/minikube/binaries/v1.28.4/kubectl ...
I0131 14:10:57.524226 125487 ssh_runner.go:362] scp memory --> /var/tmp/minikube/cni.yaml (2438 bytes)
I0131 14:10:57.612055 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl apply --kubeconfig=/var/lib/minikube/kubeconfig -f /var/tmp/minikube/cni.yaml
I0131 14:10:58.353756 125487 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0131 14:10:58.353851 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:10:58.353879 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl label nodes minikube.k8s.io/version=v1.32.0 minikube.k8s.io/commit=274d15c48919de599d1c531208ca35671bcbf218 minikube.k8s.io/name=addons-214491 minikube.k8s.io/updated_at=2024_01_31T14_10_58_0700 minikube.k8s.io/primary=true --all --overwrite --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:10:58.361795 125487 ops.go:34] apiserver oom_adj: -16
I0131 14:10:58.438468 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:10:58.939033 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:10:59.439434 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:10:59.939454 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:00.438654 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:00.938533 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:01.438556 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:01.938944 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:02.438797 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:02.939398 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:03.439413 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:03.939560 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:04.438813 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:04.939337 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:05.439313 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:05.939387 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:06.439137 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:06.939316 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:07.439395 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:07.938747 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:08.438929 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:08.939341 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:09.439141 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:09.938594 125487 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.28.4/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0131 14:11:10.013637 125487 kubeadm.go:1088] duration metric: took 11.65984962s to wait for elevateKubeSystemPrivileges.
I0131 14:11:10.013681 125487 kubeadm.go:406] StartCluster complete in 21.337849721s
I0131 14:11:10.013707 125487 settings.go:142] acquiring lock: {Name:mk5c6ffd872c98cabf7f960ae9ce352ca09ff7c1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:11:10.013847 125487 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/18007-117277/kubeconfig
I0131 14:11:10.014387 125487 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/18007-117277/kubeconfig: {Name:mkc4792adf9cf6b0aa335f112c622bcab16b5821 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0131 14:11:10.014657 125487 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.4/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0131 14:11:10.014910 125487 config.go:182] Loaded profile config "addons-214491": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.4
I0131 14:11:10.014858 125487 addons.go:502] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volumesnapshots:true yakd:true]
I0131 14:11:10.015007 125487 addons.go:69] Setting yakd=true in profile "addons-214491"
I0131 14:11:10.015031 125487 addons.go:234] Setting addon yakd=true in "addons-214491"
I0131 14:11:10.015088 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.015148 125487 addons.go:69] Setting ingress-dns=true in profile "addons-214491"
I0131 14:11:10.015178 125487 addons.go:234] Setting addon ingress-dns=true in "addons-214491"
I0131 14:11:10.015239 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.015617 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.015662 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.015841 125487 addons.go:69] Setting default-storageclass=true in profile "addons-214491"
I0131 14:11:10.015878 125487 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-214491"
I0131 14:11:10.015878 125487 addons.go:69] Setting cloud-spanner=true in profile "addons-214491"
I0131 14:11:10.015908 125487 addons.go:234] Setting addon cloud-spanner=true in "addons-214491"
I0131 14:11:10.015968 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.016117 125487 addons.go:69] Setting inspektor-gadget=true in profile "addons-214491"
I0131 14:11:10.016142 125487 addons.go:234] Setting addon inspektor-gadget=true in "addons-214491"
I0131 14:11:10.016175 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.016187 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.016449 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.016644 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.016640 125487 addons.go:69] Setting storage-provisioner=true in profile "addons-214491"
I0131 14:11:10.016669 125487 addons.go:234] Setting addon storage-provisioner=true in "addons-214491"
I0131 14:11:10.016652 125487 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-214491"
I0131 14:11:10.016678 125487 addons.go:69] Setting volumesnapshots=true in profile "addons-214491"
I0131 14:11:10.016718 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.016724 125487 addons.go:234] Setting addon volumesnapshots=true in "addons-214491"
I0131 14:11:10.016748 125487 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-214491"
I0131 14:11:10.016803 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.016805 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.016884 125487 addons.go:69] Setting gcp-auth=true in profile "addons-214491"
I0131 14:11:10.016909 125487 mustload.go:65] Loading cluster: addons-214491
I0131 14:11:10.017120 125487 config.go:182] Loaded profile config "addons-214491": Driver=docker, ContainerRuntime=containerd, KubernetesVersion=v1.28.4
I0131 14:11:10.017160 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.017262 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.017367 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.017432 125487 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-214491"
I0131 14:11:10.017453 125487 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-214491"
I0131 14:11:10.017506 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.017946 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.019161 125487 addons.go:69] Setting registry=true in profile "addons-214491"
I0131 14:11:10.019239 125487 addons.go:234] Setting addon registry=true in "addons-214491"
I0131 14:11:10.019303 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.019803 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.021521 125487 addons.go:69] Setting metrics-server=true in profile "addons-214491"
I0131 14:11:10.021615 125487 addons.go:234] Setting addon metrics-server=true in "addons-214491"
I0131 14:11:10.021713 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.022172 125487 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-214491"
I0131 14:11:10.022220 125487 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-214491"
I0131 14:11:10.022341 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.022589 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.022673 125487 addons.go:69] Setting ingress=true in profile "addons-214491"
I0131 14:11:10.022695 125487 addons.go:234] Setting addon ingress=true in "addons-214491"
I0131 14:11:10.022760 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.023021 125487 addons.go:69] Setting helm-tiller=true in profile "addons-214491"
I0131 14:11:10.023075 125487 addons.go:234] Setting addon helm-tiller=true in "addons-214491"
I0131 14:11:10.023147 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.017262 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.055440 125487 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.24.0
I0131 14:11:10.056865 125487 addons.go:426] installing /etc/kubernetes/addons/ig-namespace.yaml
I0131 14:11:10.056905 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0131 14:11:10.056986 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.055152 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.059819 125487 out.go:177] - Using image docker.io/registry:2.8.3
I0131 14:11:10.060071 125487 addons.go:234] Setting addon default-storageclass=true in "addons-214491"
I0131 14:11:10.059211 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.058838 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.068383 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.069688 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.073570 125487 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.2
I0131 14:11:10.075390 125487 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.13
I0131 14:11:10.075399 125487 addons.go:426] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0131 14:11:10.073977 125487 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.14.3
I0131 14:11:10.068403 125487 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.5
I0131 14:11:10.077166 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0131 14:11:10.077291 125487 addons.go:426] installing /etc/kubernetes/addons/deployment.yaml
I0131 14:11:10.078572 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0131 14:11:10.078665 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.081091 125487 addons.go:426] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0131 14:11:10.081120 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0131 14:11:10.081190 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.083220 125487 addons.go:426] installing /etc/kubernetes/addons/registry-rc.yaml
I0131 14:11:10.083254 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (798 bytes)
I0131 14:11:10.078990 125487 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.4
I0131 14:11:10.079059 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.083324 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.085683 125487 addons.go:426] installing /etc/kubernetes/addons/yakd-ns.yaml
I0131 14:11:10.086169 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0131 14:11:10.085826 125487 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0131 14:11:10.087641 125487 addons.go:426] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0131 14:11:10.087665 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0131 14:11:10.087740 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.086509 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.100550 125487 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0131 14:11:10.103223 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.105426 125487 addons.go:426] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0131 14:11:10.105458 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0131 14:11:10.105553 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.123467 125487 addons.go:426] installing /etc/kubernetes/addons/storageclass.yaml
I0131 14:11:10.123503 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0131 14:11:10.123571 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.132227 125487 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0131 14:11:10.132737 125487 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-214491"
I0131 14:11:10.134577 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:10.135127 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:10.135311 125487 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0131 14:11:10.144294 125487 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0131 14:11:10.146229 125487 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0131 14:11:10.147468 125487 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0131 14:11:10.148686 125487 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0131 14:11:10.150203 125487 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0131 14:11:10.156042 125487 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.6.4
I0131 14:11:10.157946 125487 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0131 14:11:10.159270 125487 addons.go:426] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0131 14:11:10.159286 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0131 14:11:10.158029 125487 addons.go:426] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0131 14:11:10.159310 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0131 14:11:10.159348 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.159359 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.158367 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.159123 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.162885 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.165154 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.168099 125487 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0131 14:11:10.169562 125487 addons.go:426] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0131 14:11:10.169587 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0131 14:11:10.169645 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.171646 125487 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20231011-8b53cabe0
I0131 14:11:10.172988 125487 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.9.5
I0131 14:11:10.174395 125487 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20231011-8b53cabe0
I0131 14:11:10.175821 125487 addons.go:426] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0131 14:11:10.175848 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16103 bytes)
I0131 14:11:10.175922 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.177592 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.181059 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.184743 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.192475 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.196313 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.202567 125487 out.go:177] - Using image docker.io/busybox:stable
I0131 14:11:10.206128 125487 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0131 14:11:10.207949 125487 addons.go:426] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0131 14:11:10.207971 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0131 14:11:10.208041 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:10.207521 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.207526 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.207521 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
W0131 14:11:10.213697 125487 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0131 14:11:10.213738 125487 retry.go:31] will retry after 176.610455ms: ssh: handshake failed: EOF
W0131 14:11:10.213775 125487 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0131 14:11:10.213799 125487 retry.go:31] will retry after 243.480794ms: ssh: handshake failed: EOF
I0131 14:11:10.243370 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:10.412745 125487 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.4/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.28.4/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0131 14:11:10.518350 125487 addons.go:426] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0131 14:11:10.518381 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0131 14:11:10.521594 125487 kapi.go:248] "coredns" deployment in "kube-system" namespace and "addons-214491" context rescaled to 1 replicas
I0131 14:11:10.521657 125487 start.go:223] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.28.4 ContainerRuntime:containerd ControlPlane:true Worker:true}
I0131 14:11:10.523165 125487 out.go:177] * Verifying Kubernetes components...
I0131 14:11:10.524922 125487 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0131 14:11:10.609386 125487 addons.go:426] installing /etc/kubernetes/addons/registry-svc.yaml
I0131 14:11:10.609426 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0131 14:11:10.610012 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0131 14:11:10.624875 125487 addons.go:426] installing /etc/kubernetes/addons/ig-role.yaml
I0131 14:11:10.624918 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0131 14:11:10.813834 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0131 14:11:10.814993 125487 addons.go:426] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0131 14:11:10.815024 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0131 14:11:10.817346 125487 addons.go:426] installing /etc/kubernetes/addons/yakd-sa.yaml
I0131 14:11:10.817375 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0131 14:11:10.819036 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0131 14:11:10.822634 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0131 14:11:10.906971 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0131 14:11:10.907320 125487 addons.go:426] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0131 14:11:10.907348 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0131 14:11:10.910049 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0131 14:11:11.009771 125487 addons.go:426] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0131 14:11:11.009808 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0131 14:11:11.011493 125487 addons.go:426] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0131 14:11:11.011526 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0131 14:11:11.024700 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0131 14:11:11.105441 125487 addons.go:426] installing /etc/kubernetes/addons/registry-proxy.yaml
I0131 14:11:11.105498 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0131 14:11:11.113400 125487 addons.go:426] installing /etc/kubernetes/addons/yakd-crb.yaml
I0131 14:11:11.113435 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0131 14:11:11.209370 125487 addons.go:426] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0131 14:11:11.209426 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0131 14:11:11.306833 125487 addons.go:426] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0131 14:11:11.306871 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0131 14:11:11.308707 125487 addons.go:426] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0131 14:11:11.308738 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0131 14:11:11.405834 125487 addons.go:426] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0131 14:11:11.405875 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0131 14:11:11.418136 125487 addons.go:426] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0131 14:11:11.418178 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0131 14:11:11.608424 125487 addons.go:426] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0131 14:11:11.608465 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0131 14:11:11.612813 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0131 14:11:11.705516 125487 addons.go:426] installing /etc/kubernetes/addons/yakd-svc.yaml
I0131 14:11:11.705552 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0131 14:11:11.721201 125487 addons.go:426] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0131 14:11:11.721237 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0131 14:11:11.822114 125487 addons.go:426] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0131 14:11:11.822149 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0131 14:11:11.924152 125487 addons.go:426] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0131 14:11:11.924190 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0131 14:11:12.010412 125487 addons.go:426] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0131 14:11:12.010449 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0131 14:11:12.206875 125487 addons.go:426] installing /etc/kubernetes/addons/ig-crd.yaml
I0131 14:11:12.206978 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0131 14:11:12.213716 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0131 14:11:12.404355 125487 addons.go:426] installing /etc/kubernetes/addons/yakd-dp.yaml
I0131 14:11:12.404395 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0131 14:11:12.407811 125487 addons.go:426] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0131 14:11:12.407847 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0131 14:11:12.408224 125487 addons.go:426] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0131 14:11:12.408252 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0131 14:11:12.412459 125487 addons.go:426] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0131 14:11:12.412490 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0131 14:11:12.423802 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0131 14:11:12.817358 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0131 14:11:12.908687 125487 addons.go:426] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0131 14:11:12.908723 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0131 14:11:12.923984 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0131 14:11:13.217168 125487 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.28.4/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.28.4/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (2.804317398s)
I0131 14:11:13.217249 125487 start.go:929] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0131 14:11:13.217285 125487 ssh_runner.go:235] Completed: sudo systemctl is-active --quiet service kubelet: (2.692337667s)
I0131 14:11:13.218417 125487 node_ready.go:35] waiting up to 6m0s for node "addons-214491" to be "Ready" ...
I0131 14:11:13.218671 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (2.608630575s)
I0131 14:11:13.222182 125487 node_ready.go:49] node "addons-214491" has status "Ready":"True"
I0131 14:11:13.222223 125487 node_ready.go:38] duration metric: took 3.773062ms waiting for node "addons-214491" to be "Ready" ...
I0131 14:11:13.222235 125487 pod_ready.go:35] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0131 14:11:13.230834 125487 pod_ready.go:78] waiting up to 6m0s for pod "coredns-5dd5756b68-dhqkg" in "kube-system" namespace to be "Ready" ...
I0131 14:11:13.303808 125487 addons.go:426] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0131 14:11:13.303854 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0131 14:11:13.324676 125487 addons.go:426] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0131 14:11:13.324790 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0131 14:11:13.715280 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0131 14:11:13.905866 125487 addons.go:426] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0131 14:11:13.905909 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0131 14:11:14.214152 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (3.400266253s)
I0131 14:11:14.706825 125487 addons.go:426] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0131 14:11:14.706867 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0131 14:11:15.221200 125487 addons.go:426] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0131 14:11:15.221311 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0131 14:11:15.404658 125487 pod_ready.go:102] pod "coredns-5dd5756b68-dhqkg" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:15.521177 125487 addons.go:426] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0131 14:11:15.521272 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0131 14:11:15.804479 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (4.985395233s)
I0131 14:11:16.113763 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0131 14:11:16.909895 125487 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0131 14:11:16.909982 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:16.942318 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:17.719427 125487 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0131 14:11:17.823180 125487 pod_ready.go:102] pod "coredns-5dd5756b68-dhqkg" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:17.919588 125487 addons.go:234] Setting addon gcp-auth=true in "addons-214491"
I0131 14:11:17.919725 125487 host.go:66] Checking if "addons-214491" exists ...
I0131 14:11:17.920472 125487 cli_runner.go:164] Run: docker container inspect addons-214491 --format={{.State.Status}}
I0131 14:11:17.940154 125487 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0131 14:11:17.940204 125487 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-214491
I0131 14:11:17.957372 125487 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32772 SSHKeyPath:/home/jenkins/minikube-integration/18007-117277/.minikube/machines/addons-214491/id_rsa Username:docker}
I0131 14:11:19.926804 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (9.104114158s)
I0131 14:11:19.926860 125487 addons.go:470] Verifying addon ingress=true in "addons-214491"
I0131 14:11:19.928798 125487 out.go:177] * Verifying ingress addon...
I0131 14:11:19.926951 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (9.016870973s)
I0131 14:11:19.927022 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (8.902295612s)
I0131 14:11:19.927046 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (9.019866766s)
I0131 14:11:19.927126 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (8.314200988s)
I0131 14:11:19.927161 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (7.713402308s)
I0131 14:11:19.927259 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (7.503420588s)
I0131 14:11:19.927305 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (7.109907374s)
I0131 14:11:19.927415 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (7.003383335s)
I0131 14:11:19.927559 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (6.212178691s)
I0131 14:11:19.928912 125487 addons.go:470] Verifying addon metrics-server=true in "addons-214491"
I0131 14:11:19.928950 125487 addons.go:470] Verifying addon registry=true in "addons-214491"
W0131 14:11:19.929002 125487 addons.go:452] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0131 14:11:19.931803 125487 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0131 14:11:19.932261 125487 out.go:177] * Verifying registry addon...
I0131 14:11:19.932287 125487 retry.go:31] will retry after 239.16958ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0131 14:11:19.933760 125487 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-214491 service yakd-dashboard -n yakd-dashboard
I0131 14:11:19.937414 125487 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0131 14:11:20.004237 125487 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0131 14:11:20.004851 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:20.010288 125487 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0131 14:11:20.010320 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:20.174101 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0131 14:11:20.305414 125487 pod_ready.go:102] pod "coredns-5dd5756b68-dhqkg" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:20.516704 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:20.521154 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:20.938321 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:21.010054 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:21.439659 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:21.514253 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:21.827161 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (5.713291803s)
I0131 14:11:21.827264 125487 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (3.887078266s)
I0131 14:11:21.827281 125487 addons.go:470] Verifying addon csi-hostpath-driver=true in "addons-214491"
I0131 14:11:21.829155 125487 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.0
I0131 14:11:21.831912 125487 out.go:177] * Verifying csi-hostpath-driver addon...
I0131 14:11:21.833365 125487 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v20231011-8b53cabe0
I0131 14:11:21.834870 125487 addons.go:426] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0131 14:11:21.834953 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0131 14:11:21.834086 125487 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0131 14:11:21.908948 125487 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0131 14:11:21.909032 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:21.928728 125487 addons.go:426] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0131 14:11:21.928759 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0131 14:11:22.003677 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:22.011624 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:22.020497 125487 addons.go:426] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0131 14:11:22.020630 125487 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5432 bytes)
I0131 14:11:22.110416 125487 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0131 14:11:22.310021 125487 pod_ready.go:102] pod "coredns-5dd5756b68-dhqkg" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:22.407269 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:22.505807 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:22.511325 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:22.818136 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.643950223s)
I0131 14:11:22.841579 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:22.939577 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:23.011473 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:23.322418 125487 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.28.4/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.211952152s)
I0131 14:11:23.323354 125487 addons.go:470] Verifying addon gcp-auth=true in "addons-214491"
I0131 14:11:23.324890 125487 out.go:177] * Verifying gcp-auth addon...
I0131 14:11:23.327175 125487 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0131 14:11:23.330516 125487 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0131 14:11:23.330545 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:23.405068 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:23.437845 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:23.510412 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:23.831448 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:23.840727 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:23.938065 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:24.010748 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:24.331757 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:24.341273 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:24.438671 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:24.510107 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:24.738479 125487 pod_ready.go:92] pod "coredns-5dd5756b68-dhqkg" in "kube-system" namespace has status "Ready":"True"
I0131 14:11:24.738505 125487 pod_ready.go:81] duration metric: took 11.50764341s waiting for pod "coredns-5dd5756b68-dhqkg" in "kube-system" namespace to be "Ready" ...
I0131 14:11:24.738515 125487 pod_ready.go:78] waiting up to 6m0s for pod "etcd-addons-214491" in "kube-system" namespace to be "Ready" ...
I0131 14:11:24.744197 125487 pod_ready.go:92] pod "etcd-addons-214491" in "kube-system" namespace has status "Ready":"True"
I0131 14:11:24.744229 125487 pod_ready.go:81] duration metric: took 5.706672ms waiting for pod "etcd-addons-214491" in "kube-system" namespace to be "Ready" ...
I0131 14:11:24.744248 125487 pod_ready.go:78] waiting up to 6m0s for pod "kube-apiserver-addons-214491" in "kube-system" namespace to be "Ready" ...
I0131 14:11:24.752331 125487 pod_ready.go:92] pod "kube-apiserver-addons-214491" in "kube-system" namespace has status "Ready":"True"
I0131 14:11:24.752356 125487 pod_ready.go:81] duration metric: took 8.099299ms waiting for pod "kube-apiserver-addons-214491" in "kube-system" namespace to be "Ready" ...
I0131 14:11:24.752368 125487 pod_ready.go:78] waiting up to 6m0s for pod "kube-controller-manager-addons-214491" in "kube-system" namespace to be "Ready" ...
I0131 14:11:24.757550 125487 pod_ready.go:92] pod "kube-controller-manager-addons-214491" in "kube-system" namespace has status "Ready":"True"
I0131 14:11:24.757578 125487 pod_ready.go:81] duration metric: took 5.202559ms waiting for pod "kube-controller-manager-addons-214491" in "kube-system" namespace to be "Ready" ...
I0131 14:11:24.757592 125487 pod_ready.go:78] waiting up to 6m0s for pod "kube-proxy-6sbxl" in "kube-system" namespace to be "Ready" ...
I0131 14:11:24.762934 125487 pod_ready.go:92] pod "kube-proxy-6sbxl" in "kube-system" namespace has status "Ready":"True"
I0131 14:11:24.762958 125487 pod_ready.go:81] duration metric: took 5.357685ms waiting for pod "kube-proxy-6sbxl" in "kube-system" namespace to be "Ready" ...
I0131 14:11:24.762971 125487 pod_ready.go:78] waiting up to 6m0s for pod "kube-scheduler-addons-214491" in "kube-system" namespace to be "Ready" ...
I0131 14:11:24.831677 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:24.841739 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:24.938561 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:25.010463 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:25.135292 125487 pod_ready.go:92] pod "kube-scheduler-addons-214491" in "kube-system" namespace has status "Ready":"True"
I0131 14:11:25.135324 125487 pod_ready.go:81] duration metric: took 372.344643ms waiting for pod "kube-scheduler-addons-214491" in "kube-system" namespace to be "Ready" ...
I0131 14:11:25.135339 125487 pod_ready.go:78] waiting up to 6m0s for pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace to be "Ready" ...
I0131 14:11:25.330757 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:25.340024 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:25.438991 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:25.510140 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:25.832044 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:25.841063 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:25.938221 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:26.010464 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:26.331927 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:26.340811 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:26.438997 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:26.509881 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:26.831692 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:26.840782 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:26.938652 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:27.009797 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:27.140655 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:27.330655 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:27.339814 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:27.438030 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:27.510506 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:27.831249 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:27.841864 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:27.939332 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:28.010077 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:28.331957 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:28.341273 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:28.438498 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:28.510374 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:28.831118 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:28.840155 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:28.938072 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:29.009530 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:29.141402 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:29.330957 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:29.341291 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:29.438579 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:29.511036 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:29.831841 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:29.841844 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:29.938776 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:30.010630 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:30.331697 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:30.342942 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:30.437420 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:30.509994 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:30.831333 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:30.842008 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:30.939880 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:31.010029 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:31.144159 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:31.331977 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:31.341058 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:31.438727 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:31.510245 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:31.832126 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:31.841393 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:31.940218 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:32.010203 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:32.336994 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:32.341363 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:32.438447 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:32.510779 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:32.831083 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:32.840911 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:32.937729 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:33.009250 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0131 14:11:33.331141 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:33.340665 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:33.441676 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:33.509708 125487 kapi.go:107] duration metric: took 13.505475467s to wait for kubernetes.io/minikube-addons=registry ...
I0131 14:11:33.642842 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:33.831804 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:33.841714 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:33.938603 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:34.331125 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:34.345205 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:34.438100 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:34.832124 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:34.841294 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:34.938454 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:35.331141 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:35.342090 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:35.438541 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:35.831586 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:35.840976 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:35.938244 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:36.141622 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:36.331217 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:36.342275 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:36.438800 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:36.831622 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:36.840858 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:36.940018 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:37.331290 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:37.341292 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:37.438447 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:37.830801 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:37.840000 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:37.938815 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:38.142608 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:38.332114 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:38.341055 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:38.438357 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:38.832357 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:38.841881 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:38.940524 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:39.331332 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:39.341053 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:39.439305 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:39.831435 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:39.842091 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:39.938518 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:40.142701 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:40.331594 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:40.340714 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:40.439983 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:40.831785 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:40.842014 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:40.939288 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:41.331693 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:41.341185 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:41.439465 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:41.832325 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:41.841415 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:41.939384 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:42.332421 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:42.342724 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:42.438399 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:42.641764 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:42.831103 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:42.840276 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:42.938306 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:43.331284 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:43.340568 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:43.451455 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:43.832286 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:43.841845 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:43.939576 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:44.331466 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:44.340303 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:44.480026 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:44.641798 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:44.873133 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:44.875354 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:45.055446 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:45.335001 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:45.340682 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:45.438333 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:45.831442 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:45.840768 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:45.941026 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:46.330975 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:46.340700 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:46.438894 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:46.700293 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:46.831187 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:46.840073 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:46.937952 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:47.332605 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:47.341267 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:47.441980 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:47.831326 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:47.840675 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:47.938403 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:48.332226 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:48.341964 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:48.438346 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:48.831208 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:48.841575 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:48.939781 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:49.142798 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:49.331572 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:49.340679 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:49.440839 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:49.831075 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:49.840937 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:49.938405 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:50.332091 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:50.340549 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:50.439558 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:50.831914 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:50.866960 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:50.938385 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:51.331934 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:51.340903 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:51.438209 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:51.641154 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:51.831348 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:51.840904 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:51.939358 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:52.332119 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:52.341719 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:52.439853 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:52.832984 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:52.842095 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:52.939349 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:53.331393 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:53.341021 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:53.438693 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:53.641976 125487 pod_ready.go:102] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"False"
I0131 14:11:53.832060 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:53.839739 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:53.938070 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:54.330667 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:54.341865 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:54.438798 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:54.641811 125487 pod_ready.go:92] pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace has status "Ready":"True"
I0131 14:11:54.641834 125487 pod_ready.go:81] duration metric: took 29.506487947s waiting for pod "nvidia-device-plugin-daemonset-lmr4m" in "kube-system" namespace to be "Ready" ...
I0131 14:11:54.641843 125487 pod_ready.go:38] duration metric: took 41.419592825s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0131 14:11:54.641861 125487 api_server.go:52] waiting for apiserver process to appear ...
I0131 14:11:54.641906 125487 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0131 14:11:54.658004 125487 api_server.go:72] duration metric: took 44.136299685s to wait for apiserver process to appear ...
I0131 14:11:54.658036 125487 api_server.go:88] waiting for apiserver healthz status ...
I0131 14:11:54.658059 125487 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0131 14:11:54.663491 125487 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0131 14:11:54.664767 125487 api_server.go:141] control plane version: v1.28.4
I0131 14:11:54.664789 125487 api_server.go:131] duration metric: took 6.747154ms to wait for apiserver health ...
I0131 14:11:54.664798 125487 system_pods.go:43] waiting for kube-system pods to appear ...
I0131 14:11:54.675007 125487 system_pods.go:59] 19 kube-system pods found
I0131 14:11:54.675038 125487 system_pods.go:61] "coredns-5dd5756b68-dhqkg" [4ee3da37-ee93-4adf-84b5-1a3536d0affd] Running
I0131 14:11:54.675043 125487 system_pods.go:61] "csi-hostpath-attacher-0" [88bb39b3-bbd4-43e7-82ab-e15251e315b8] Running
I0131 14:11:54.675048 125487 system_pods.go:61] "csi-hostpath-resizer-0" [60f6c692-f7d8-4bd5-a237-bae55df662c8] Running
I0131 14:11:54.675056 125487 system_pods.go:61] "csi-hostpathplugin-kwkf2" [f15a28c1-2b0e-4ffb-97c2-1998273acef4] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0131 14:11:54.675072 125487 system_pods.go:61] "etcd-addons-214491" [c18cf44a-a75f-4461-a983-3a75fd339fd7] Running
I0131 14:11:54.675078 125487 system_pods.go:61] "kindnet-p9gwl" [efb988f3-3908-4431-a07e-99d45208a690] Running
I0131 14:11:54.675082 125487 system_pods.go:61] "kube-apiserver-addons-214491" [8bd186fc-2746-445f-9c00-133f576943f6] Running
I0131 14:11:54.675087 125487 system_pods.go:61] "kube-controller-manager-addons-214491" [76e0cfb3-278c-4699-9ff6-a78134d6209a] Running
I0131 14:11:54.675091 125487 system_pods.go:61] "kube-ingress-dns-minikube" [9c551d17-79a7-4334-94f8-60c40742d004] Running
I0131 14:11:54.675095 125487 system_pods.go:61] "kube-proxy-6sbxl" [d50a831b-a3d1-4061-8e3c-89413d6438dd] Running
I0131 14:11:54.675099 125487 system_pods.go:61] "kube-scheduler-addons-214491" [7158bd24-bf9b-4f68-995e-8f49346ea88f] Running
I0131 14:11:54.675105 125487 system_pods.go:61] "metrics-server-7c66d45ddc-5z5sm" [f00d08cb-8ef5-4fb1-9a0f-bc55ce02a581] Running
I0131 14:11:54.675112 125487 system_pods.go:61] "nvidia-device-plugin-daemonset-lmr4m" [3c951f22-d962-4f13-929a-e7a2552f629c] Running
I0131 14:11:54.675116 125487 system_pods.go:61] "registry-d9s2q" [2871b25a-9352-469f-8a23-944ab9a8e387] Running
I0131 14:11:54.675123 125487 system_pods.go:61] "registry-proxy-q74f8" [df7b6b84-753f-4801-9602-60eb8519d1b6] Running
I0131 14:11:54.675127 125487 system_pods.go:61] "snapshot-controller-58dbcc7b99-4rngn" [3cb8e201-d1cf-4d76-9062-94fad1c64649] Running
I0131 14:11:54.675134 125487 system_pods.go:61] "snapshot-controller-58dbcc7b99-8xhng" [ac6a57e5-7da4-4d0e-86ce-7bf9f6c20fc2] Running
I0131 14:11:54.675138 125487 system_pods.go:61] "storage-provisioner" [58e6a917-440c-442d-866f-9fb81149f70f] Running
I0131 14:11:54.675144 125487 system_pods.go:61] "tiller-deploy-7b677967b9-fpdpc" [df50ab8f-7dde-4c15-ac50-a9d5d3ee508e] Running
I0131 14:11:54.675152 125487 system_pods.go:74] duration metric: took 10.348002ms to wait for pod list to return data ...
I0131 14:11:54.675162 125487 default_sa.go:34] waiting for default service account to be created ...
I0131 14:11:54.677286 125487 default_sa.go:45] found service account: "default"
I0131 14:11:54.677306 125487 default_sa.go:55] duration metric: took 2.138733ms for default service account to be created ...
I0131 14:11:54.677313 125487 system_pods.go:116] waiting for k8s-apps to be running ...
I0131 14:11:54.685019 125487 system_pods.go:86] 19 kube-system pods found
I0131 14:11:54.685048 125487 system_pods.go:89] "coredns-5dd5756b68-dhqkg" [4ee3da37-ee93-4adf-84b5-1a3536d0affd] Running
I0131 14:11:54.685054 125487 system_pods.go:89] "csi-hostpath-attacher-0" [88bb39b3-bbd4-43e7-82ab-e15251e315b8] Running
I0131 14:11:54.685058 125487 system_pods.go:89] "csi-hostpath-resizer-0" [60f6c692-f7d8-4bd5-a237-bae55df662c8] Running
I0131 14:11:54.685065 125487 system_pods.go:89] "csi-hostpathplugin-kwkf2" [f15a28c1-2b0e-4ffb-97c2-1998273acef4] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0131 14:11:54.685070 125487 system_pods.go:89] "etcd-addons-214491" [c18cf44a-a75f-4461-a983-3a75fd339fd7] Running
I0131 14:11:54.685077 125487 system_pods.go:89] "kindnet-p9gwl" [efb988f3-3908-4431-a07e-99d45208a690] Running
I0131 14:11:54.685085 125487 system_pods.go:89] "kube-apiserver-addons-214491" [8bd186fc-2746-445f-9c00-133f576943f6] Running
I0131 14:11:54.685089 125487 system_pods.go:89] "kube-controller-manager-addons-214491" [76e0cfb3-278c-4699-9ff6-a78134d6209a] Running
I0131 14:11:54.685100 125487 system_pods.go:89] "kube-ingress-dns-minikube" [9c551d17-79a7-4334-94f8-60c40742d004] Running
I0131 14:11:54.685105 125487 system_pods.go:89] "kube-proxy-6sbxl" [d50a831b-a3d1-4061-8e3c-89413d6438dd] Running
I0131 14:11:54.685109 125487 system_pods.go:89] "kube-scheduler-addons-214491" [7158bd24-bf9b-4f68-995e-8f49346ea88f] Running
I0131 14:11:54.685116 125487 system_pods.go:89] "metrics-server-7c66d45ddc-5z5sm" [f00d08cb-8ef5-4fb1-9a0f-bc55ce02a581] Running
I0131 14:11:54.685121 125487 system_pods.go:89] "nvidia-device-plugin-daemonset-lmr4m" [3c951f22-d962-4f13-929a-e7a2552f629c] Running
I0131 14:11:54.685126 125487 system_pods.go:89] "registry-d9s2q" [2871b25a-9352-469f-8a23-944ab9a8e387] Running
I0131 14:11:54.685130 125487 system_pods.go:89] "registry-proxy-q74f8" [df7b6b84-753f-4801-9602-60eb8519d1b6] Running
I0131 14:11:54.685135 125487 system_pods.go:89] "snapshot-controller-58dbcc7b99-4rngn" [3cb8e201-d1cf-4d76-9062-94fad1c64649] Running
I0131 14:11:54.685139 125487 system_pods.go:89] "snapshot-controller-58dbcc7b99-8xhng" [ac6a57e5-7da4-4d0e-86ce-7bf9f6c20fc2] Running
I0131 14:11:54.685143 125487 system_pods.go:89] "storage-provisioner" [58e6a917-440c-442d-866f-9fb81149f70f] Running
I0131 14:11:54.685146 125487 system_pods.go:89] "tiller-deploy-7b677967b9-fpdpc" [df50ab8f-7dde-4c15-ac50-a9d5d3ee508e] Running
I0131 14:11:54.685152 125487 system_pods.go:126] duration metric: took 7.833768ms to wait for k8s-apps to be running ...
I0131 14:11:54.685159 125487 system_svc.go:44] waiting for kubelet service to be running ....
I0131 14:11:54.685214 125487 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0131 14:11:54.696577 125487 system_svc.go:56] duration metric: took 11.410467ms WaitForService to wait for kubelet.
I0131 14:11:54.696598 125487 kubeadm.go:581] duration metric: took 44.174901213s to wait for : map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] ...
I0131 14:11:54.696618 125487 node_conditions.go:102] verifying NodePressure condition ...
I0131 14:11:54.699284 125487 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0131 14:11:54.699310 125487 node_conditions.go:123] node cpu capacity is 8
I0131 14:11:54.699323 125487 node_conditions.go:105] duration metric: took 2.701422ms to run NodePressure ...
I0131 14:11:54.699332 125487 start.go:228] waiting for startup goroutines ...
I0131 14:11:54.831496 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:54.840276 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:54.938365 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:55.331861 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:55.341301 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:55.439986 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:55.832641 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:55.840481 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:55.938810 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:56.331076 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:56.340879 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:56.438956 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:56.831061 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:56.842210 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:56.938343 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:57.331854 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:57.342766 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:57.439382 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:57.832216 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:57.842069 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:57.938269 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:58.331640 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:58.340721 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:58.481705 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:58.832298 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:58.841891 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:58.938671 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:59.332401 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:59.342436 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:59.439024 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:11:59.831264 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:11:59.841828 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:11:59.939857 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:12:00.332232 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:12:00.341898 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:00.439564 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:12:00.832026 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:12:00.842410 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:00.938373 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:12:01.331104 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:12:01.340556 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:01.438916 125487 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0131 14:12:01.831662 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:12:01.841262 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:01.938945 125487 kapi.go:107] duration metric: took 42.007141412s to wait for app.kubernetes.io/name=ingress-nginx ...
I0131 14:12:02.332002 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0131 14:12:02.341332 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:02.835639 125487 kapi.go:107] duration metric: took 39.508461558s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0131 14:12:02.837932 125487 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-214491 cluster.
I0131 14:12:02.839475 125487 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0131 14:12:02.840944 125487 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0131 14:12:02.842883 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:03.340650 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:03.841045 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:04.341161 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:04.843041 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:05.340225 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:05.841901 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:06.340700 125487 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0131 14:12:06.840313 125487 kapi.go:107] duration metric: took 45.006225231s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0131 14:12:06.842220 125487 out.go:177] * Enabled addons: nvidia-device-plugin, ingress-dns, storage-provisioner-rancher, storage-provisioner, inspektor-gadget, cloud-spanner, helm-tiller, metrics-server, yakd, default-storageclass, volumesnapshots, registry, ingress, gcp-auth, csi-hostpath-driver
I0131 14:12:06.843506 125487 addons.go:505] enable addons completed in 56.828654914s: enabled=[nvidia-device-plugin ingress-dns storage-provisioner-rancher storage-provisioner inspektor-gadget cloud-spanner helm-tiller metrics-server yakd default-storageclass volumesnapshots registry ingress gcp-auth csi-hostpath-driver]
I0131 14:12:06.843557 125487 start.go:233] waiting for cluster config update ...
I0131 14:12:06.843579 125487 start.go:242] writing updated cluster config ...
I0131 14:12:06.843860 125487 ssh_runner.go:195] Run: rm -f paused
I0131 14:12:06.895175 125487 start.go:600] kubectl: 1.29.1, cluster: 1.28.4 (minor skew: 1)
I0131 14:12:06.897172 125487 out.go:177] * Done! kubectl is now configured to use "addons-214491" cluster and "default" namespace by default
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
20dec4e4614d7 2b70e4aaac6b5 5 seconds ago Running nginx 0 b13f4340cd1ae nginx
ff5683c1bfec2 98f6c3b32d565 14 seconds ago Exited helm-test 0 c29b3cc573346 helm-test
98029f7d670dc 738351fd438f0 24 seconds ago Running csi-snapshotter 0 6bbbeaef7e618 csi-hostpathplugin-kwkf2
b440280239fdf 931dbfd16f87c 25 seconds ago Running csi-provisioner 0 6bbbeaef7e618 csi-hostpathplugin-kwkf2
e1a7670d4586d e899260153aed 26 seconds ago Running liveness-probe 0 6bbbeaef7e618 csi-hostpathplugin-kwkf2
f8bbdbf3c9286 e255e073c508c 26 seconds ago Running hostpath 0 6bbbeaef7e618 csi-hostpathplugin-kwkf2
b7b65d3658d6a 6d2a98b274382 27 seconds ago Running gcp-auth 0 6f68dd99f8994 gcp-auth-d4c87556c-fdqtx
2cf2b7fee92e2 311f90a3747fd 28 seconds ago Running controller 0 184155fe0f85c ingress-nginx-controller-69cff4fd79-6l2mk
a09fd8cb278b8 88ef14a257f42 33 seconds ago Running node-driver-registrar 0 6bbbeaef7e618 csi-hostpathplugin-kwkf2
8e47ffaa61098 754854eab8c1c 34 seconds ago Running cloud-spanner-emulator 0 9be687c468e30 cloud-spanner-emulator-64c8c85f65-44b5j
3dfcb37066e22 8cfc3f994a82b 36 seconds ago Running nvidia-device-plugin-ctr 0 379f8c333b784 nvidia-device-plugin-daemonset-lmr4m
d563e8cea6cd1 1ebff0f9671bc 40 seconds ago Exited patch 0 d0249dd47af3e gcp-auth-certs-patch-m7mv6
8b2dc0799c98c 1ebff0f9671bc 41 seconds ago Exited create 0 9ff9e172f3fec gcp-auth-certs-create-kxrtn
866f317e0e8ff 19a639eda60f0 41 seconds ago Running csi-resizer 0 7b88ff114de0f csi-hostpath-resizer-0
09775a461131c a1ed5895ba635 42 seconds ago Running csi-external-health-monitor-controller 0 6bbbeaef7e618 csi-hostpathplugin-kwkf2
c782fabfb6f86 59cbb42146a37 43 seconds ago Running csi-attacher 0 bcdfa8e527f98 csi-hostpath-attacher-0
59991ecd7a71c 1ebff0f9671bc 43 seconds ago Exited patch 1 8224220cf630e ingress-nginx-admission-patch-8xvd6
bb5d5599012d8 1ebff0f9671bc 44 seconds ago Exited create 0 b4b46f313ee77 ingress-nginx-admission-create-rcgr8
1eb30f1d5c4e6 aa61ee9c70bc4 47 seconds ago Running volume-snapshot-controller 0 311e2fa269427 snapshot-controller-58dbcc7b99-4rngn
22ceb40395172 aa61ee9c70bc4 47 seconds ago Running volume-snapshot-controller 0 06b569d7fa873 snapshot-controller-58dbcc7b99-8xhng
d457d8c6bcf4b 31de47c733c91 52 seconds ago Running yakd 0 791ad26f8fafa yakd-dashboard-9947fc6bf-tjtlr
65e7dc3df70cb e16d1e3a10667 59 seconds ago Running local-path-provisioner 0 ca740c0b5150d local-path-provisioner-78b46b4d5c-vxxxf
11c4ec48778f5 1499ed4fbd0aa About a minute ago Running minikube-ingress-dns 0 08257abf5dd5c kube-ingress-dns-minikube
20082807b22ba ead0a4a53df89 About a minute ago Running coredns 0 ba86e99709d10 coredns-5dd5756b68-dhqkg
027c3c119c2c0 6e38f40d628db About a minute ago Running storage-provisioner 0 6566cadc9df3d storage-provisioner
fc5708a499819 c7d1297425461 About a minute ago Running kindnet-cni 0 0d267afdf0390 kindnet-p9gwl
2ada3aafb372e 83f6cc407eed8 About a minute ago Running kube-proxy 0 db86f64fcab06 kube-proxy-6sbxl
1795212d12e86 d058aa5ab969c About a minute ago Running kube-controller-manager 0 b0ef8c3a18072 kube-controller-manager-addons-214491
abb617918ca21 e3db313c6dbc0 About a minute ago Running kube-scheduler 0 99fa2a58612bb kube-scheduler-addons-214491
22e2442985946 73deb9a3f7025 About a minute ago Running etcd 0 e5982d8b897ed etcd-addons-214491
0eb1d87425a8e 7fe0e6f37db33 About a minute ago Running kube-apiserver 0 f9ce0afcccc00 kube-apiserver-addons-214491
==> containerd <==
Jan 31 14:12:25 addons-214491 containerd[783]: time="2024-01-31T14:12:25.532195647Z" level=info msg="PullImage \"docker.io/busybox:stable@sha256:3fbc632167424a6d997e74f52b878d7cc478225cffac6bc977eedfe51c7f4e79\""
Jan 31 14:12:25 addons-214491 containerd[783]: time="2024-01-31T14:12:25.534011487Z" level=error msg="failed to decode hosts.toml" error="invalid `host` tree"
Jan 31 14:12:25 addons-214491 containerd[783]: time="2024-01-31T14:12:25.813995982Z" level=error msg="failed to decode hosts.toml" error="invalid `host` tree"
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.563819342Z" level=info msg="ImageCreate event &ImageCreate{Name:docker.io/library/busybox@sha256:3fbc632167424a6d997e74f52b878d7cc478225cffac6bc977eedfe51c7f4e79,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}"
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.566379293Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:a416a98b71e224a31ee99cff8e16063554498227d2b696152a9c3e0aa65e5824,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}"
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.568257455Z" level=info msg="ImageUpdate event &ImageUpdate{Name:docker.io/library/busybox@sha256:3fbc632167424a6d997e74f52b878d7cc478225cffac6bc977eedfe51c7f4e79,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}"
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.568724836Z" level=info msg="PullImage \"docker.io/busybox:stable@sha256:3fbc632167424a6d997e74f52b878d7cc478225cffac6bc977eedfe51c7f4e79\" returns image reference \"sha256:a416a98b71e224a31ee99cff8e16063554498227d2b696152a9c3e0aa65e5824\""
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.570698649Z" level=info msg="CreateContainer within sandbox \"3ff841a74152d483fd0dc29276ab3069cdd544cc0daaab6dfe5c219480f3e9aa\" for container &ContainerMetadata{Name:helper-pod,Attempt:0,}"
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.616245556Z" level=info msg="CreateContainer within sandbox \"3ff841a74152d483fd0dc29276ab3069cdd544cc0daaab6dfe5c219480f3e9aa\" for &ContainerMetadata{Name:helper-pod,Attempt:0,} returns container id \"884b7e4b417699eabfd37cd0a8f6064205665f2bbac9e47fc438a94470964d77\""
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.616961905Z" level=info msg="StartContainer for \"884b7e4b417699eabfd37cd0a8f6064205665f2bbac9e47fc438a94470964d77\""
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.667113596Z" level=info msg="StartContainer for \"884b7e4b417699eabfd37cd0a8f6064205665f2bbac9e47fc438a94470964d77\" returns successfully"
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.775017279Z" level=info msg="shim disconnected" id=884b7e4b417699eabfd37cd0a8f6064205665f2bbac9e47fc438a94470964d77
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.775091696Z" level=warning msg="cleaning up after shim disconnected" id=884b7e4b417699eabfd37cd0a8f6064205665f2bbac9e47fc438a94470964d77 namespace=k8s.io
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.775108473Z" level=info msg="cleaning up dead shim"
Jan 31 14:12:26 addons-214491 containerd[783]: time="2024-01-31T14:12:26.783759242Z" level=warning msg="cleanup warnings time=\"2024-01-31T14:12:26Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=8352 runtime=io.containerd.runc.v2\n"
Jan 31 14:12:28 addons-214491 containerd[783]: time="2024-01-31T14:12:28.742755197Z" level=info msg="StopPodSandbox for \"3ff841a74152d483fd0dc29276ab3069cdd544cc0daaab6dfe5c219480f3e9aa\""
Jan 31 14:12:28 addons-214491 containerd[783]: time="2024-01-31T14:12:28.742853511Z" level=info msg="Container to stop \"884b7e4b417699eabfd37cd0a8f6064205665f2bbac9e47fc438a94470964d77\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jan 31 14:12:28 addons-214491 containerd[783]: time="2024-01-31T14:12:28.773238286Z" level=info msg="shim disconnected" id=3ff841a74152d483fd0dc29276ab3069cdd544cc0daaab6dfe5c219480f3e9aa
Jan 31 14:12:28 addons-214491 containerd[783]: time="2024-01-31T14:12:28.773531510Z" level=warning msg="cleaning up after shim disconnected" id=3ff841a74152d483fd0dc29276ab3069cdd544cc0daaab6dfe5c219480f3e9aa namespace=k8s.io
Jan 31 14:12:28 addons-214491 containerd[783]: time="2024-01-31T14:12:28.773560823Z" level=info msg="cleaning up dead shim"
Jan 31 14:12:28 addons-214491 containerd[783]: time="2024-01-31T14:12:28.782721072Z" level=warning msg="cleanup warnings time=\"2024-01-31T14:12:28Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=8406 runtime=io.containerd.runc.v2\n"
Jan 31 14:12:28 addons-214491 containerd[783]: time="2024-01-31T14:12:28.826638769Z" level=info msg="TearDown network for sandbox \"3ff841a74152d483fd0dc29276ab3069cdd544cc0daaab6dfe5c219480f3e9aa\" successfully"
Jan 31 14:12:28 addons-214491 containerd[783]: time="2024-01-31T14:12:28.826683245Z" level=info msg="StopPodSandbox for \"3ff841a74152d483fd0dc29276ab3069cdd544cc0daaab6dfe5c219480f3e9aa\" returns successfully"
Jan 31 14:12:29 addons-214491 containerd[783]: time="2024-01-31T14:12:29.748160041Z" level=info msg="RemoveContainer for \"884b7e4b417699eabfd37cd0a8f6064205665f2bbac9e47fc438a94470964d77\""
Jan 31 14:12:29 addons-214491 containerd[783]: time="2024-01-31T14:12:29.753458664Z" level=info msg="RemoveContainer for \"884b7e4b417699eabfd37cd0a8f6064205665f2bbac9e47fc438a94470964d77\" returns successfully"
==> coredns [20082807b22bafba9d25817d55fbf255dd277544471bc4bc2728d4222a6b8b37] <==
[INFO] 10.244.0.3:52200 - 47021 "A IN registry.kube-system.svc.cluster.local.cluster.local. udp 70 false 512" NXDOMAIN qr,aa,rd 163 0.000098574s
[INFO] 10.244.0.3:33863 - 61967 "A IN registry.kube-system.svc.cluster.local.us-central1-a.c.k8s-minikube.internal. udp 94 false 512" NXDOMAIN qr,rd,ra 94 0.006507s
[INFO] 10.244.0.3:33863 - 60928 "AAAA IN registry.kube-system.svc.cluster.local.us-central1-a.c.k8s-minikube.internal. udp 94 false 512" NXDOMAIN qr,rd,ra 94 0.008150932s
[INFO] 10.244.0.3:51781 - 3670 "AAAA IN registry.kube-system.svc.cluster.local.c.k8s-minikube.internal. udp 80 false 512" NXDOMAIN qr,rd,ra 80 0.006176873s
[INFO] 10.244.0.3:51781 - 26451 "A IN registry.kube-system.svc.cluster.local.c.k8s-minikube.internal. udp 80 false 512" NXDOMAIN qr,rd,ra 80 0.006255234s
[INFO] 10.244.0.3:59764 - 9735 "A IN registry.kube-system.svc.cluster.local.google.internal. udp 72 false 512" NXDOMAIN qr,rd,ra 72 0.006394931s
[INFO] 10.244.0.3:59764 - 29048 "AAAA IN registry.kube-system.svc.cluster.local.google.internal. udp 72 false 512" NXDOMAIN qr,rd,ra 72 0.006565105s
[INFO] 10.244.0.3:57305 - 14744 "AAAA IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 149 0.000074065s
[INFO] 10.244.0.3:57305 - 50837 "A IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 110 0.000129805s
[INFO] 10.244.0.21:56913 - 2570 "AAAA IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000213681s
[INFO] 10.244.0.21:42105 - 7099 "A IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000353481s
[INFO] 10.244.0.21:55828 - 52802 "A IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000152557s
[INFO] 10.244.0.21:34693 - 63415 "AAAA IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000128043s
[INFO] 10.244.0.21:41318 - 10247 "AAAA IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000107131s
[INFO] 10.244.0.21:44480 - 46445 "A IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000125879s
[INFO] 10.244.0.21:52879 - 19020 "AAAA IN storage.googleapis.com.us-central1-a.c.k8s-minikube.internal. udp 89 false 1232" NXDOMAIN qr,rd,ra 78 0.009336902s
[INFO] 10.244.0.21:39852 - 22586 "A IN storage.googleapis.com.us-central1-a.c.k8s-minikube.internal. udp 89 false 1232" NXDOMAIN qr,rd,ra 78 0.01415022s
[INFO] 10.244.0.21:49406 - 52533 "AAAA IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.008342825s
[INFO] 10.244.0.21:60202 - 12859 "A IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.010133937s
[INFO] 10.244.0.21:39803 - 8702 "A IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.009125735s
[INFO] 10.244.0.21:51045 - 48296 "AAAA IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.024088416s
[INFO] 10.244.0.21:44937 - 60599 "AAAA IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 240 0.001348958s
[INFO] 10.244.0.21:49016 - 55787 "A IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 648 0.001447792s
[INFO] 10.244.0.23:42892 - 2 "AAAA IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 149 0.000177132s
[INFO] 10.244.0.23:39712 - 3 "A IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 110 0.000171085s
==> describe nodes <==
Name: addons-214491
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-214491
kubernetes.io/os=linux
minikube.k8s.io/commit=274d15c48919de599d1c531208ca35671bcbf218
minikube.k8s.io/name=addons-214491
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_01_31T14_10_58_0700
minikube.k8s.io/version=v1.32.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-214491
Annotations: csi.volume.kubernetes.io/nodeid: {"hostpath.csi.k8s.io":"addons-214491"}
kubeadm.alpha.kubernetes.io/cri-socket: unix:///run/containerd/containerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Wed, 31 Jan 2024 14:10:54 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-214491
AcquireTime: <unset>
RenewTime: Wed, 31 Jan 2024 14:12:29 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Wed, 31 Jan 2024 14:12:29 +0000 Wed, 31 Jan 2024 14:10:52 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Wed, 31 Jan 2024 14:12:29 +0000 Wed, 31 Jan 2024 14:10:52 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Wed, 31 Jan 2024 14:12:29 +0000 Wed, 31 Jan 2024 14:10:52 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Wed, 31 Jan 2024 14:12:29 +0000 Wed, 31 Jan 2024 14:10:57 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-214491
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859428Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859428Ki
pods: 110
System Info:
Machine ID: 1513117ec6364ff3ab682486b7669ee9
System UUID: ad3e0fa0-48f4-451b-a763-e0e2854f4f71
Boot ID: 59de63ea-ec4c-4e26-a911-a59699678b11
Kernel Version: 5.15.0-1049-gcp
OS Image: Ubuntu 22.04.3 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: containerd://1.6.26
Kubelet Version: v1.28.4
Kube-Proxy Version: v1.28.4
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (21 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default cloud-spanner-emulator-64c8c85f65-44b5j 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 77s
default nginx 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 8s
gcp-auth gcp-auth-d4c87556c-fdqtx 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 67s
ingress-nginx ingress-nginx-controller-69cff4fd79-6l2mk 100m (1%!)(MISSING) 0 (0%!)(MISSING) 90Mi (0%!)(MISSING) 0 (0%!)(MISSING) 71s
kube-system coredns-5dd5756b68-dhqkg 100m (1%!)(MISSING) 0 (0%!)(MISSING) 70Mi (0%!)(MISSING) 170Mi (0%!)(MISSING) 80s
kube-system csi-hostpath-attacher-0 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 69s
kube-system csi-hostpath-resizer-0 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 69s
kube-system csi-hostpathplugin-kwkf2 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 69s
kube-system etcd-addons-214491 100m (1%!)(MISSING) 0 (0%!)(MISSING) 100Mi (0%!)(MISSING) 0 (0%!)(MISSING) 93s
kube-system kindnet-p9gwl 100m (1%!)(MISSING) 100m (1%!)(MISSING) 50Mi (0%!)(MISSING) 50Mi (0%!)(MISSING) 81s
kube-system kube-apiserver-addons-214491 250m (3%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 93s
kube-system kube-controller-manager-addons-214491 200m (2%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 93s
kube-system kube-ingress-dns-minikube 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 76s
kube-system kube-proxy-6sbxl 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 81s
kube-system kube-scheduler-addons-214491 100m (1%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 93s
kube-system nvidia-device-plugin-daemonset-lmr4m 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 78s
kube-system snapshot-controller-58dbcc7b99-4rngn 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 72s
kube-system snapshot-controller-58dbcc7b99-8xhng 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 72s
kube-system storage-provisioner 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 74s
local-path-storage local-path-provisioner-78b46b4d5c-vxxxf 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 75s
yakd-dashboard yakd-dashboard-9947fc6bf-tjtlr 0 (0%!)(MISSING) 0 (0%!)(MISSING) 128Mi (0%!)(MISSING) 256Mi (0%!)(MISSING) 73s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 950m (11%!)(MISSING) 100m (1%!)(MISSING)
memory 438Mi (1%!)(MISSING) 476Mi (1%!)(MISSING)
ephemeral-storage 0 (0%!)(MISSING) 0 (0%!)(MISSING)
hugepages-1Gi 0 (0%!)(MISSING) 0 (0%!)(MISSING)
hugepages-2Mi 0 (0%!)(MISSING) 0 (0%!)(MISSING)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 77s kube-proxy
Normal Starting 99s kubelet Starting kubelet.
Normal NodeHasSufficientMemory 99s (x8 over 99s) kubelet Node addons-214491 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 99s (x8 over 99s) kubelet Node addons-214491 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 99s (x7 over 99s) kubelet Node addons-214491 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 99s kubelet Updated Node Allocatable limit across pods
Normal Starting 93s kubelet Starting kubelet.
Normal NodeHasSufficientMemory 93s kubelet Node addons-214491 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 93s kubelet Node addons-214491 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 93s kubelet Node addons-214491 status is now: NodeHasSufficientPID
Normal NodeNotReady 93s kubelet Node addons-214491 status is now: NodeNotReady
Normal NodeAllocatableEnforced 93s kubelet Updated Node Allocatable limit across pods
Normal NodeReady 93s kubelet Node addons-214491 status is now: NodeReady
Normal RegisteredNode 81s node-controller Node addons-214491 event: Registered Node addons-214491 in Controller
==> dmesg <==
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 2e ce 7e e2 57 50 08 06
[Jan31 13:47] IPv4: martian source 10.244.0.1 from 10.244.0.29, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 66 7d f5 31 e3 29 08 06
[ +0.000158] IPv4: martian source 10.244.0.1 from 10.244.0.28, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 36 03 8f a6 69 94 08 06
[ +19.874195] IPv4: martian source 10.244.0.1 from 10.244.0.30, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff e2 8f 4a 31 49 10 08 06
[Jan31 13:48] IPv4: martian source 10.244.0.1 from 10.244.0.31, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 1a f1 65 e4 70 47 08 06
[ +25.959432] IPv4: martian source 10.244.0.1 from 10.244.0.34, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 7e 7f 9c 50 2d 37 08 06
[Jan31 13:50] IPv4: martian source 10.244.0.1 from 10.244.0.40, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff ba be d3 f7 c8 ec 08 06
[Jan31 13:52] IPv4: martian source 10.244.0.1 from 10.244.0.41, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 2a 86 8e 5c f6 ff 08 06
[ +15.266675] IPv4: martian source 10.244.0.1 from 10.244.0.42, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff e6 d0 4b b7 1d f3 08 06
[Jan31 13:53] IPv4: martian source 10.244.0.1 from 10.244.0.43, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 9a 68 39 f5 41 71 08 06
[Jan31 13:54] IPv4: martian source 10.244.0.1 from 10.244.0.44, on dev eth0
[ +0.000013] ll header: 00000000: ff ff ff ff ff ff f2 c9 46 5b 03 bb 08 06
[ +32.842842] IPv4: martian source 10.244.0.1 from 10.244.0.2, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff be 8d 2c 5f db d3 08 06
[Jan31 13:55] IPv4: martian source 10.244.0.1 from 10.244.0.3, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff fe 96 a9 b7 6c 7d 08 06
==> etcd [22e244298594624880ed12b9a118863444f627bae00137f11d2c1b5a6e759ff8] <==
{"level":"info","ts":"2024-01-31T14:10:52.126361Z","caller":"embed/etcd.go:855","msg":"serving metrics","address":"http://127.0.0.1:2381"}
{"level":"info","ts":"2024-01-31T14:10:52.126498Z","caller":"embed/etcd.go:597","msg":"serving peer traffic","address":"192.168.49.2:2380"}
{"level":"info","ts":"2024-01-31T14:10:52.126517Z","caller":"embed/etcd.go:569","msg":"cmux::serve","address":"192.168.49.2:2380"}
{"level":"info","ts":"2024-01-31T14:10:52.214956Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc is starting a new election at term 1"}
{"level":"info","ts":"2024-01-31T14:10:52.215086Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became pre-candidate at term 1"}
{"level":"info","ts":"2024-01-31T14:10:52.215142Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgPreVoteResp from aec36adc501070cc at term 1"}
{"level":"info","ts":"2024-01-31T14:10:52.215229Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became candidate at term 2"}
{"level":"info","ts":"2024-01-31T14:10:52.215261Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-01-31T14:10:52.215307Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-01-31T14:10:52.21535Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-01-31T14:10:52.216289Z","caller":"etcdserver/server.go:2062","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-214491 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-01-31T14:10:52.216328Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-01-31T14:10:52.216384Z","caller":"etcdserver/server.go:2571","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-01-31T14:10:52.216599Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-01-31T14:10:52.216669Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-01-31T14:10:52.216355Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-01-31T14:10:52.217056Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-01-31T14:10:52.217237Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-01-31T14:10:52.217305Z","caller":"etcdserver/server.go:2595","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-01-31T14:10:52.217775Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-01-31T14:10:52.218017Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-01-31T14:11:43.617313Z","caller":"traceutil/trace.go:171","msg":"trace[1345552921] transaction","detail":"{read_only:false; response_revision:992; number_of_response:1; }","duration":"155.011457ms","start":"2024-01-31T14:11:43.462264Z","end":"2024-01-31T14:11:43.617276Z","steps":["trace[1345552921] 'process raft request' (duration: 110.796561ms)","trace[1345552921] 'compare' (duration: 44.063544ms)"],"step_count":2}
{"level":"warn","ts":"2024-01-31T14:11:45.052966Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"116.828794ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods/ingress-nginx/\" range_end:\"/registry/pods/ingress-nginx0\" ","response":"range_response_count:3 size:13482"}
{"level":"info","ts":"2024-01-31T14:11:45.053091Z","caller":"traceutil/trace.go:171","msg":"trace[30449622] range","detail":"{range_begin:/registry/pods/ingress-nginx/; range_end:/registry/pods/ingress-nginx0; response_count:3; response_revision:997; }","duration":"116.972323ms","start":"2024-01-31T14:11:44.936095Z","end":"2024-01-31T14:11:45.053067Z","steps":["trace[30449622] 'range keys from in-memory index tree' (duration: 116.690517ms)"],"step_count":1}
{"level":"info","ts":"2024-01-31T14:12:15.613985Z","caller":"traceutil/trace.go:171","msg":"trace[1935600369] transaction","detail":"{read_only:false; response_revision:1230; number_of_response:1; }","duration":"106.973517ms","start":"2024-01-31T14:12:15.506981Z","end":"2024-01-31T14:12:15.613954Z","steps":["trace[1935600369] 'process raft request' (duration: 106.787826ms)"],"step_count":1}
==> gcp-auth [b7b65d3658d6ae0158e48347ab670d1af878312fa5c528eb8578344af056a040] <==
2024/01/31 14:12:02 GCP Auth Webhook started!
2024/01/31 14:12:13 Ready to marshal response ...
2024/01/31 14:12:13 Ready to write response ...
2024/01/31 14:12:18 Ready to marshal response ...
2024/01/31 14:12:18 Ready to write response ...
2024/01/31 14:12:22 Ready to marshal response ...
2024/01/31 14:12:22 Ready to write response ...
2024/01/31 14:12:25 Ready to marshal response ...
2024/01/31 14:12:25 Ready to write response ...
2024/01/31 14:12:25 Ready to marshal response ...
2024/01/31 14:12:25 Ready to write response ...
==> kernel <==
14:12:30 up 18:55, 0 users, load average: 1.35, 0.87, 0.67
Linux addons-214491 5.15.0-1049-gcp #57~20.04.1-Ubuntu SMP Wed Jan 17 16:04:23 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.3 LTS"
==> kindnet [fc5708a4998190033321ebad158cf42da467aef8369c4dc1afdc709839f7ae5f] <==
I0131 14:11:11.708869 1 main.go:102] connected to apiserver: https://10.96.0.1:443
I0131 14:11:11.708952 1 main.go:107] hostIP = 192.168.49.2
podIP = 192.168.49.2
I0131 14:11:11.709113 1 main.go:116] setting mtu 1500 for CNI
I0131 14:11:11.709124 1 main.go:146] kindnetd IP family: "ipv4"
I0131 14:11:11.709147 1 main.go:150] noMask IPv4 subnets: [10.244.0.0/16]
I0131 14:11:12.104434 1 main.go:191] Failed to get nodes, retrying after error: Get "https://10.96.0.1:443/api/v1/nodes": dial tcp 10.96.0.1:443: connect: connection refused
I0131 14:11:12.104842 1 main.go:191] Failed to get nodes, retrying after error: Get "https://10.96.0.1:443/api/v1/nodes": dial tcp 10.96.0.1:443: connect: connection refused
I0131 14:11:13.111729 1 main.go:191] Failed to get nodes, retrying after error: Get "https://10.96.0.1:443/api/v1/nodes": dial tcp 10.96.0.1:443: connect: connection refused
I0131 14:11:15.226704 1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
I0131 14:11:15.226743 1 main.go:227] handling current node
I0131 14:11:25.324928 1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
I0131 14:11:25.324956 1 main.go:227] handling current node
I0131 14:11:35.339072 1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
I0131 14:11:35.339177 1 main.go:227] handling current node
I0131 14:11:45.342907 1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
I0131 14:11:45.342934 1 main.go:227] handling current node
I0131 14:11:55.353381 1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
I0131 14:11:55.353421 1 main.go:227] handling current node
I0131 14:12:05.365601 1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
I0131 14:12:05.365631 1 main.go:227] handling current node
I0131 14:12:15.370353 1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
I0131 14:12:15.370384 1 main.go:227] handling current node
I0131 14:12:25.374071 1 main.go:223] Handling node with IPs: map[192.168.49.2:{}]
I0131 14:12:25.374100 1 main.go:227] handling current node
==> kube-apiserver [0eb1d87425a8ea62511f1efd17c5f5a4e11c3559c05612f9e10cdfa8c233b54a] <==
W0131 14:11:35.352570 1 handler_proxy.go:93] no RequestInfo found in the context
E0131 14:11:35.352603 1 available_controller.go:460] v1beta1.metrics.k8s.io failed with: failing or missing response from https://10.110.208.117:443/apis/metrics.k8s.io/v1beta1: Get "https://10.110.208.117:443/apis/metrics.k8s.io/v1beta1": dial tcp 10.110.208.117:443: connect: connection refused
E0131 14:11:35.352727 1 controller.go:146] Error updating APIService "v1beta1.metrics.k8s.io" with err: failed to download v1beta1.metrics.k8s.io: failed to retrieve openAPI spec, http error: ResponseCode: 503, Body: service unavailable
, Header: map[Content-Type:[text/plain; charset=utf-8] X-Content-Type-Options:[nosniff]]
I0131 14:11:35.406465 1 handler.go:232] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
W0131 14:11:36.355277 1 handler_proxy.go:93] no RequestInfo found in the context
W0131 14:11:36.355292 1 handler_proxy.go:93] no RequestInfo found in the context
E0131 14:11:36.355313 1 controller.go:113] loading OpenAPI spec for "v1beta1.metrics.k8s.io" failed with: Error, could not get list of group versions for APIService
I0131 14:11:36.355322 1 controller.go:126] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Rate Limited Requeue.
E0131 14:11:36.355391 1 controller.go:102] loading OpenAPI spec for "v1beta1.metrics.k8s.io" failed with: failed to download v1beta1.metrics.k8s.io: failed to retrieve openAPI spec, http error: ResponseCode: 503, Body: service unavailable
, Header: map[Content-Type:[text/plain; charset=utf-8] X-Content-Type-Options:[nosniff]]
I0131 14:11:36.356353 1 controller.go:109] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Rate Limited Requeue.
W0131 14:11:40.359819 1 handler_proxy.go:93] no RequestInfo found in the context
E0131 14:11:40.359887 1 controller.go:146] Error updating APIService "v1beta1.metrics.k8s.io" with err: failed to download v1beta1.metrics.k8s.io: failed to retrieve openAPI spec, http error: ResponseCode: 503, Body: service unavailable
, Header: map[Content-Type:[text/plain; charset=utf-8] X-Content-Type-Options:[nosniff]]
E0131 14:11:40.360079 1 available_controller.go:460] v1beta1.metrics.k8s.io failed with: failing or missing response from https://10.110.208.117:443/apis/metrics.k8s.io/v1beta1: Get "https://10.110.208.117:443/apis/metrics.k8s.io/v1beta1": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
I0131 14:11:40.365369 1 handler.go:232] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
I0131 14:11:40.419543 1 handler.go:232] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
I0131 14:11:54.162571 1 handler.go:232] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager
I0131 14:12:19.418293 1 handler.go:232] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
I0131 14:12:19.424456 1 handler.go:232] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0131 14:12:20.434755 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
E0131 14:12:20.750776 1 watch.go:287] unable to encode watch object *v1.WatchEvent: http2: stream closed (&streaming.encoderWithAllocator{writer:responsewriter.outerWithCloseNotifyAndFlush{UserProvidedDecorator:(*metrics.ResponseWriterDelegator)(0xc00a16db60), InnerCloseNotifierFlusher:struct { httpsnoop.Unwrapper; http.ResponseWriter; http.Flusher; http.CloseNotifier; http.Pusher }{Unwrapper:(*httpsnoop.rw)(0xc008338910), ResponseWriter:(*httpsnoop.rw)(0xc008338910), Flusher:(*httpsnoop.rw)(0xc008338910), CloseNotifier:(*httpsnoop.rw)(0xc008338910), Pusher:(*httpsnoop.rw)(0xc008338910)}}, encoder:(*versioning.codec)(0xc00a4ba280), memAllocator:(*runtime.Allocator)(0xc00d24d410)})
I0131 14:12:21.956222 1 controller.go:624] quota admission added evaluator for: ingresses.networking.k8s.io
I0131 14:12:22.149358 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.101.206.180"}
==> kube-controller-manager [1795212d12e8678f8bfce160f7b6fabd45d51e877117df973fd7994e1f1a1ddd] <==
I0131 14:12:02.541816 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="gcp-auth/gcp-auth-d4c87556c" duration="7.636236ms"
I0131 14:12:02.541961 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="gcp-auth/gcp-auth-d4c87556c" duration="85.677µs"
I0131 14:12:07.092247 1 event.go:307] "Event occurred" object="default/hpvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
I0131 14:12:09.057404 1 event.go:307] "Event occurred" object="default/hpvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
I0131 14:12:11.940495 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="ingress-nginx/ingress-nginx-controller-69cff4fd79" duration="8.808827ms"
I0131 14:12:11.941385 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="ingress-nginx/ingress-nginx-controller-69cff4fd79" duration="95.766µs"
I0131 14:12:13.493964 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/metrics-server-7c66d45ddc" duration="7.021µs"
I0131 14:12:18.364208 1 replica_set.go:676] "Finished syncing" kind="ReplicaSet" key="kube-system/tiller-deploy-7b677967b9" duration="9.343µs"
I0131 14:12:18.406370 1 event.go:307] "Event occurred" object="kube-system/tiller-deploy" fieldPath="" kind="Endpoints" apiVersion="v1" type="Warning" reason="FailedToUpdateEndpoint" message="Failed to update endpoint kube-system/tiller-deploy: Operation cannot be fulfilled on endpoints \"tiller-deploy\": StorageError: invalid object, Code: 4, Key: /registry/services/endpoints/kube-system/tiller-deploy, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 05bdba54-4641-4bea-afef-e91530a56dea, UID in object meta: "
E0131 14:12:20.436413 1 reflector.go:147] vendor/k8s.io/client-go/metadata/metadatainformer/informer.go:106: Failed to watch *v1.PartialObjectMetadata: the server could not find the requested resource
I0131 14:12:21.015843 1 job_controller.go:562] "enqueueing job" key="gcp-auth/gcp-auth-certs-patch"
I0131 14:12:21.017198 1 job_controller.go:562] "enqueueing job" key="gcp-auth/gcp-auth-certs-create"
I0131 14:12:21.036477 1 job_controller.go:562] "enqueueing job" key="gcp-auth/gcp-auth-certs-patch"
I0131 14:12:21.037571 1 job_controller.go:562] "enqueueing job" key="gcp-auth/gcp-auth-certs-create"
I0131 14:12:21.459311 1 replica_set.go:676] "Finished syncing" kind="ReplicationController" key="kube-system/registry" duration="15.058µs"
W0131 14:12:21.703557 1 reflector.go:535] vendor/k8s.io/client-go/metadata/metadatainformer/informer.go:106: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0131 14:12:21.703634 1 reflector.go:147] vendor/k8s.io/client-go/metadata/metadatainformer/informer.go:106: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
I0131 14:12:24.058528 1 event.go:307] "Event occurred" object="default/hpvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'hostpath.csi.k8s.io' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
W0131 14:12:24.759469 1 reflector.go:535] vendor/k8s.io/client-go/metadata/metadatainformer/informer.go:106: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0131 14:12:24.759507 1 reflector.go:147] vendor/k8s.io/client-go/metadata/metadatainformer/informer.go:106: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
I0131 14:12:24.979768 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="WaitForFirstConsumer" message="waiting for first consumer to be created before binding"
I0131 14:12:25.111909 1 event.go:307] "Event occurred" object="default/test-pvc" fieldPath="" kind="PersistentVolumeClaim" apiVersion="v1" type="Normal" reason="ExternalProvisioning" message="Waiting for a volume to be created either by the external provisioner 'rancher.io/local-path' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered."
I0131 14:12:29.712225 1 namespace_controller.go:182] "Namespace has been deleted" namespace="gadget"
W0131 14:12:30.323855 1 reflector.go:535] vendor/k8s.io/client-go/metadata/metadatainformer/informer.go:106: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0131 14:12:30.323889 1 reflector.go:147] vendor/k8s.io/client-go/metadata/metadatainformer/informer.go:106: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
==> kube-proxy [2ada3aafb372ef92f7ed3b75d17e6990c046d6ef832cafc13b03c52e09d1e3f7] <==
I0131 14:11:11.805583 1 server_others.go:69] "Using iptables proxy"
I0131 14:11:11.917858 1 node.go:141] Successfully retrieved node IP: 192.168.49.2
I0131 14:11:12.211978 1 server.go:632] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0131 14:11:12.220459 1 server_others.go:152] "Using iptables Proxier"
I0131 14:11:12.220525 1 server_others.go:421] "Detect-local-mode set to ClusterCIDR, but no cluster CIDR for family" ipFamily="IPv6"
I0131 14:11:12.220536 1 server_others.go:438] "Defaulting to no-op detect-local"
I0131 14:11:12.220578 1 proxier.go:251] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses"
I0131 14:11:12.220822 1 server.go:846] "Version info" version="v1.28.4"
I0131 14:11:12.220840 1 server.go:848] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0131 14:11:12.222554 1 config.go:315] "Starting node config controller"
I0131 14:11:12.222576 1 shared_informer.go:311] Waiting for caches to sync for node config
I0131 14:11:12.223048 1 config.go:97] "Starting endpoint slice config controller"
I0131 14:11:12.223060 1 shared_informer.go:311] Waiting for caches to sync for endpoint slice config
I0131 14:11:12.223117 1 config.go:188] "Starting service config controller"
I0131 14:11:12.223125 1 shared_informer.go:311] Waiting for caches to sync for service config
I0131 14:11:12.322684 1 shared_informer.go:318] Caches are synced for node config
I0131 14:11:12.323259 1 shared_informer.go:318] Caches are synced for service config
I0131 14:11:12.323329 1 shared_informer.go:318] Caches are synced for endpoint slice config
==> kube-scheduler [abb617918ca21e0b96e4ea9bf1ce7286aeef594af7b565eca149885aa576b8ba] <==
W0131 14:10:54.325580 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
W0131 14:10:54.325598 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0131 14:10:54.325619 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0131 14:10:54.325622 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
W0131 14:10:54.325416 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0131 14:10:54.325725 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
W0131 14:10:54.325491 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0131 14:10:54.325760 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
W0131 14:10:54.325591 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0131 14:10:54.325791 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
W0131 14:10:55.160691 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0131 14:10:55.160742 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
W0131 14:10:55.172176 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0131 14:10:55.172219 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
W0131 14:10:55.182558 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0131 14:10:55.182598 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
W0131 14:10:55.202075 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0131 14:10:55.202113 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
W0131 14:10:55.329504 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0131 14:10:55.329549 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
W0131 14:10:55.335904 1 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0131 14:10:55.335946 1 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
W0131 14:10:55.585561 1 reflector.go:535] pkg/server/dynamiccertificates/configmap_cafile_content.go:206: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0131 14:10:55.585596 1 reflector.go:147] pkg/server/dynamiccertificates/configmap_cafile_content.go:206: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
I0131 14:10:57.521696 1 shared_informer.go:318] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Jan 31 14:12:25 addons-214491 kubelet[1508]: I0131 14:12:25.155819 1508 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/host-path/3b0d3c22-62d4-414d-ae65-43aa549800af-data\") pod \"helper-pod-create-pvc-b6ca553a-3812-43ca-8a9b-b1a71b4e3891\" (UID: \"3b0d3c22-62d4-414d-ae65-43aa549800af\") " pod="local-path-storage/helper-pod-create-pvc-b6ca553a-3812-43ca-8a9b-b1a71b4e3891"
Jan 31 14:12:25 addons-214491 kubelet[1508]: I0131 14:12:25.155858 1508 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/3b0d3c22-62d4-414d-ae65-43aa549800af-gcp-creds\") pod \"helper-pod-create-pvc-b6ca553a-3812-43ca-8a9b-b1a71b4e3891\" (UID: \"3b0d3c22-62d4-414d-ae65-43aa549800af\") " pod="local-path-storage/helper-pod-create-pvc-b6ca553a-3812-43ca-8a9b-b1a71b4e3891"
Jan 31 14:12:25 addons-214491 kubelet[1508]: I0131 14:12:25.156114 1508 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"script\" (UniqueName: \"kubernetes.io/configmap/3b0d3c22-62d4-414d-ae65-43aa549800af-script\") pod \"helper-pod-create-pvc-b6ca553a-3812-43ca-8a9b-b1a71b4e3891\" (UID: \"3b0d3c22-62d4-414d-ae65-43aa549800af\") " pod="local-path-storage/helper-pod-create-pvc-b6ca553a-3812-43ca-8a9b-b1a71b4e3891"
Jan 31 14:12:25 addons-214491 kubelet[1508]: I0131 14:12:25.740829 1508 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nginx" podStartSLOduration=1.51822477 podCreationTimestamp="2024-01-31 14:12:22 +0000 UTC" firstStartedPulling="2024-01-31 14:12:22.491202258 +0000 UTC m=+85.205105030" lastFinishedPulling="2024-01-31 14:12:24.713742803 +0000 UTC m=+87.427645567" observedRunningTime="2024-01-31 14:12:25.74050183 +0000 UTC m=+88.454404606" watchObservedRunningTime="2024-01-31 14:12:25.740765307 +0000 UTC m=+88.454668083"
Jan 31 14:12:26 addons-214491 kubelet[1508]: I0131 14:12:26.755831 1508 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="local-path-storage/helper-pod-create-pvc-b6ca553a-3812-43ca-8a9b-b1a71b4e3891" podStartSLOduration=0.71854612 podCreationTimestamp="2024-01-31 14:12:25 +0000 UTC" firstStartedPulling="2024-01-31 14:12:25.531831618 +0000 UTC m=+88.245734376" lastFinishedPulling="2024-01-31 14:12:26.569075702 +0000 UTC m=+89.282978470" observedRunningTime="2024-01-31 14:12:26.754745407 +0000 UTC m=+89.468648182" watchObservedRunningTime="2024-01-31 14:12:26.755790214 +0000 UTC m=+89.469692990"
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.880275 1508 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"data\" (UniqueName: \"kubernetes.io/host-path/3b0d3c22-62d4-414d-ae65-43aa549800af-data\") pod \"3b0d3c22-62d4-414d-ae65-43aa549800af\" (UID: \"3b0d3c22-62d4-414d-ae65-43aa549800af\") "
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.880366 1508 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhvhq\" (UniqueName: \"kubernetes.io/projected/3b0d3c22-62d4-414d-ae65-43aa549800af-kube-api-access-zhvhq\") pod \"3b0d3c22-62d4-414d-ae65-43aa549800af\" (UID: \"3b0d3c22-62d4-414d-ae65-43aa549800af\") "
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.880416 1508 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"script\" (UniqueName: \"kubernetes.io/configmap/3b0d3c22-62d4-414d-ae65-43aa549800af-script\") pod \"3b0d3c22-62d4-414d-ae65-43aa549800af\" (UID: \"3b0d3c22-62d4-414d-ae65-43aa549800af\") "
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.880446 1508 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/3b0d3c22-62d4-414d-ae65-43aa549800af-gcp-creds\") pod \"3b0d3c22-62d4-414d-ae65-43aa549800af\" (UID: \"3b0d3c22-62d4-414d-ae65-43aa549800af\") "
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.880471 1508 operation_generator.go:882] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b0d3c22-62d4-414d-ae65-43aa549800af-data" (OuterVolumeSpecName: "data") pod "3b0d3c22-62d4-414d-ae65-43aa549800af" (UID: "3b0d3c22-62d4-414d-ae65-43aa549800af"). InnerVolumeSpecName "data". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.880537 1508 operation_generator.go:882] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b0d3c22-62d4-414d-ae65-43aa549800af-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "3b0d3c22-62d4-414d-ae65-43aa549800af" (UID: "3b0d3c22-62d4-414d-ae65-43aa549800af"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.880653 1508 reconciler_common.go:300] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/3b0d3c22-62d4-414d-ae65-43aa549800af-gcp-creds\") on node \"addons-214491\" DevicePath \"\""
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.880682 1508 reconciler_common.go:300] "Volume detached for volume \"data\" (UniqueName: \"kubernetes.io/host-path/3b0d3c22-62d4-414d-ae65-43aa549800af-data\") on node \"addons-214491\" DevicePath \"\""
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.880985 1508 operation_generator.go:882] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b0d3c22-62d4-414d-ae65-43aa549800af-script" (OuterVolumeSpecName: "script") pod "3b0d3c22-62d4-414d-ae65-43aa549800af" (UID: "3b0d3c22-62d4-414d-ae65-43aa549800af"). InnerVolumeSpecName "script". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.882885 1508 operation_generator.go:882] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b0d3c22-62d4-414d-ae65-43aa549800af-kube-api-access-zhvhq" (OuterVolumeSpecName: "kube-api-access-zhvhq") pod "3b0d3c22-62d4-414d-ae65-43aa549800af" (UID: "3b0d3c22-62d4-414d-ae65-43aa549800af"). InnerVolumeSpecName "kube-api-access-zhvhq". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.981676 1508 reconciler_common.go:300] "Volume detached for volume \"script\" (UniqueName: \"kubernetes.io/configmap/3b0d3c22-62d4-414d-ae65-43aa549800af-script\") on node \"addons-214491\" DevicePath \"\""
Jan 31 14:12:28 addons-214491 kubelet[1508]: I0131 14:12:28.981720 1508 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-zhvhq\" (UniqueName: \"kubernetes.io/projected/3b0d3c22-62d4-414d-ae65-43aa549800af-kube-api-access-zhvhq\") on node \"addons-214491\" DevicePath \"\""
Jan 31 14:12:29 addons-214491 kubelet[1508]: I0131 14:12:29.517800 1508 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="3b0d3c22-62d4-414d-ae65-43aa549800af" path="/var/lib/kubelet/pods/3b0d3c22-62d4-414d-ae65-43aa549800af/volumes"
Jan 31 14:12:29 addons-214491 kubelet[1508]: I0131 14:12:29.746874 1508 scope.go:117] "RemoveContainer" containerID="884b7e4b417699eabfd37cd0a8f6064205665f2bbac9e47fc438a94470964d77"
Jan 31 14:12:30 addons-214491 kubelet[1508]: I0131 14:12:30.116907 1508 topology_manager.go:215] "Topology Admit Handler" podUID="6c549690-d1fd-44de-8f12-06862bd4372a" podNamespace="default" podName="test-local-path"
Jan 31 14:12:30 addons-214491 kubelet[1508]: E0131 14:12:30.117029 1508 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3b0d3c22-62d4-414d-ae65-43aa549800af" containerName="helper-pod"
Jan 31 14:12:30 addons-214491 kubelet[1508]: I0131 14:12:30.117095 1508 memory_manager.go:346] "RemoveStaleState removing state" podUID="3b0d3c22-62d4-414d-ae65-43aa549800af" containerName="helper-pod"
Jan 31 14:12:30 addons-214491 kubelet[1508]: I0131 14:12:30.192756 1508 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/6c549690-d1fd-44de-8f12-06862bd4372a-gcp-creds\") pod \"test-local-path\" (UID: \"6c549690-d1fd-44de-8f12-06862bd4372a\") " pod="default/test-local-path"
Jan 31 14:12:30 addons-214491 kubelet[1508]: I0131 14:12:30.192817 1508 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b6ca553a-3812-43ca-8a9b-b1a71b4e3891\" (UniqueName: \"kubernetes.io/host-path/6c549690-d1fd-44de-8f12-06862bd4372a-pvc-b6ca553a-3812-43ca-8a9b-b1a71b4e3891\") pod \"test-local-path\" (UID: \"6c549690-d1fd-44de-8f12-06862bd4372a\") " pod="default/test-local-path"
Jan 31 14:12:30 addons-214491 kubelet[1508]: I0131 14:12:30.192844 1508 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqljk\" (UniqueName: \"kubernetes.io/projected/6c549690-d1fd-44de-8f12-06862bd4372a-kube-api-access-pqljk\") pod \"test-local-path\" (UID: \"6c549690-d1fd-44de-8f12-06862bd4372a\") " pod="default/test-local-path"
==> storage-provisioner [027c3c119c2c083619acb12c682b3f1139a5133921532a56a1638c22742a6fd1] <==
I0131 14:11:18.929217 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0131 14:11:19.013761 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0131 14:11:19.013835 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0131 14:11:19.026742 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0131 14:11:19.102007 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-214491_f2f0b446-e24a-464c-acd3-36314cee216c!
I0131 14:11:19.102473 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"926895b8-0fd9-449e-8373-bf4c601c180d", APIVersion:"v1", ResourceVersion:"696", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-214491_f2f0b446-e24a-464c-acd3-36314cee216c became leader
I0131 14:11:19.202526 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-214491_f2f0b446-e24a-464c-acd3-36314cee216c!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-214491 -n addons-214491
helpers_test.go:261: (dbg) Run: kubectl --context addons-214491 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: test-local-path ingress-nginx-admission-create-rcgr8 ingress-nginx-admission-patch-8xvd6
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/NvidiaDevicePlugin]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-214491 describe pod test-local-path ingress-nginx-admission-create-rcgr8 ingress-nginx-admission-patch-8xvd6
helpers_test.go:277: (dbg) Non-zero exit: kubectl --context addons-214491 describe pod test-local-path ingress-nginx-admission-create-rcgr8 ingress-nginx-admission-patch-8xvd6: exit status 1 (71.382246ms)
-- stdout --
Name: test-local-path
Namespace: default
Priority: 0
Service Account: default
Node: addons-214491/192.168.49.2
Start Time: Wed, 31 Jan 2024 14:12:30 +0000
Labels: run=test-local-path
Annotations: <none>
Status: Pending
IP:
IPs: <none>
Containers:
busybox:
Container ID:
Image: busybox:stable
Image ID:
Port: <none>
Host Port: <none>
Command:
sh
-c
echo 'local-path-provisioner' > /test/file1
State: Waiting
Reason: ContainerCreating
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/test from data (rw)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-pqljk (ro)
Conditions:
Type Status
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
data:
Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace)
ClaimName: test-pvc
ReadOnly: false
kube-api-access-pqljk:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 1s default-scheduler Successfully assigned default/test-local-path to addons-214491
Normal Pulling 1s kubelet Pulling image "busybox:stable"
-- /stdout --
** stderr **
Error from server (NotFound): pods "ingress-nginx-admission-create-rcgr8" not found
Error from server (NotFound): pods "ingress-nginx-admission-patch-8xvd6" not found
** /stderr **
helpers_test.go:279: kubectl --context addons-214491 describe pod test-local-path ingress-nginx-admission-create-rcgr8 ingress-nginx-admission-patch-8xvd6: exit status 1
--- FAIL: TestAddons/parallel/NvidiaDevicePlugin (7.54s)