=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:332: registry stabilized in 2.021461ms
addons_test.go:334: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-85p89" [7ddd4a6c-0bb9-4cdd-b2c2-6a358cc36131] Running
addons_test.go:334: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 6.002422376s
addons_test.go:337: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-lrwnn" [727ad348-b4a0-40a9-a423-cac288b38182] Running
addons_test.go:337: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.003645936s
addons_test.go:342: (dbg) Run: kubectl --context addons-924081 delete po -l run=registry-test --now
addons_test.go:347: (dbg) Run: kubectl --context addons-924081 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:347: (dbg) Non-zero exit: kubectl --context addons-924081 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.07668605s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:349: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-924081 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:353: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:361: (dbg) Run: out/minikube-linux-amd64 -p addons-924081 ip
2024/09/15 18:10:30 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:390: (dbg) Run: out/minikube-linux-amd64 -p addons-924081 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-924081
helpers_test.go:235: (dbg) docker inspect addons-924081:
-- stdout --
[
{
"Id": "a7cb3ed838c674e83e689853d4795f1939ee1245471508e2b1ba6731e1e9be10",
"Created": "2024-09-15T17:57:25.705521635Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 20059,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-15T17:57:25.832384703Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:20d492278eed428d119466f58713403332b5d2ac1db7c6863f797e2406f2b671",
"ResolvConfPath": "/var/lib/docker/containers/a7cb3ed838c674e83e689853d4795f1939ee1245471508e2b1ba6731e1e9be10/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/a7cb3ed838c674e83e689853d4795f1939ee1245471508e2b1ba6731e1e9be10/hostname",
"HostsPath": "/var/lib/docker/containers/a7cb3ed838c674e83e689853d4795f1939ee1245471508e2b1ba6731e1e9be10/hosts",
"LogPath": "/var/lib/docker/containers/a7cb3ed838c674e83e689853d4795f1939ee1245471508e2b1ba6731e1e9be10/a7cb3ed838c674e83e689853d4795f1939ee1245471508e2b1ba6731e1e9be10-json.log",
"Name": "/addons-924081",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-924081:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-924081",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/d2a0e525c4e6f452f92853fadd7761a978cce1938f86d89fd0a33cbc212950bc-init/diff:/var/lib/docker/overlay2/98b43be93661840522f6675504552b2073bca744c9d1abb04e8ebf1b5d0c4763/diff",
"MergedDir": "/var/lib/docker/overlay2/d2a0e525c4e6f452f92853fadd7761a978cce1938f86d89fd0a33cbc212950bc/merged",
"UpperDir": "/var/lib/docker/overlay2/d2a0e525c4e6f452f92853fadd7761a978cce1938f86d89fd0a33cbc212950bc/diff",
"WorkDir": "/var/lib/docker/overlay2/d2a0e525c4e6f452f92853fadd7761a978cce1938f86d89fd0a33cbc212950bc/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "volume",
"Name": "addons-924081",
"Source": "/var/lib/docker/volumes/addons-924081/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
},
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
}
],
"Config": {
"Hostname": "addons-924081",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-924081",
"name.minikube.sigs.k8s.io": "addons-924081",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "542f46370b0447adce501b41bd00ee0b59fcd50cbeaa695b53af39b56bbe6f8a",
"SandboxKey": "/var/run/docker/netns/542f46370b04",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-924081": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "e9a913bd60604fddd0465b2a2773a4dbe96b68b8889e384fffda11e3908b4879",
"EndpointID": "866004ebbc4d5475d2448e02162ddb9c7bf43ac752c9e65d8488858df1555be4",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-924081",
"a7cb3ed838c6"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-924081 -n addons-924081
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-924081 logs -n 25
helpers_test.go:247: (dbg) Done: out/minikube-linux-amd64 -p addons-924081 logs -n 25: (1.326281315s)
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| start | --download-only -p | download-docker-579642 | jenkins | v1.34.0 | 15 Sep 24 17:57 UTC | |
| | download-docker-579642 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p download-docker-579642 | download-docker-579642 | jenkins | v1.34.0 | 15 Sep 24 17:57 UTC | 15 Sep 24 17:57 UTC |
| start | --download-only -p | binary-mirror-308651 | jenkins | v1.34.0 | 15 Sep 24 17:57 UTC | |
| | binary-mirror-308651 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:33109 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-308651 | binary-mirror-308651 | jenkins | v1.34.0 | 15 Sep 24 17:57 UTC | 15 Sep 24 17:57 UTC |
| addons | disable dashboard -p | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 17:57 UTC | |
| | addons-924081 | | | | | |
| addons | enable dashboard -p | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 17:57 UTC | |
| | addons-924081 | | | | | |
| start | -p addons-924081 --wait=true | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 17:57 UTC | 15 Sep 24 18:00 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | addons-924081 addons disable | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:01 UTC | 15 Sep 24 18:01 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | addons-924081 addons disable | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:09 UTC | 15 Sep 24 18:09 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| ssh | addons-924081 ssh cat | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:09 UTC | 15 Sep 24 18:09 UTC |
| | /opt/local-path-provisioner/pvc-c592a443-dc12-4138-ba5c-46e5f18ad12e_default_test-pvc/file1 | | | | | |
| addons | addons-924081 addons disable | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:09 UTC | 15 Sep 24 18:10 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:09 UTC | 15 Sep 24 18:09 UTC |
| | -p addons-924081 | | | | | |
| addons | disable cloud-spanner -p | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:09 UTC | 15 Sep 24 18:09 UTC |
| | addons-924081 | | | | | |
| addons | addons-924081 addons | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:09 UTC | 15 Sep 24 18:09 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-924081 addons | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:09 UTC | 15 Sep 24 18:09 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-924081 addons disable | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:09 UTC | 15 Sep 24 18:09 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-924081 addons | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:10 UTC | 15 Sep 24 18:10 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | enable headlamp | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:10 UTC | 15 Sep 24 18:10 UTC |
| | -p addons-924081 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:10 UTC | 15 Sep 24 18:10 UTC |
| | addons-924081 | | | | | |
| addons | addons-924081 addons disable | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:10 UTC | 15 Sep 24 18:10 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ssh | addons-924081 ssh curl -s | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:10 UTC | 15 Sep 24 18:10 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-924081 ip | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:10 UTC | 15 Sep 24 18:10 UTC |
| addons | addons-924081 addons disable | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:10 UTC | 15 Sep 24 18:10 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ip | addons-924081 ip | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:10 UTC | 15 Sep 24 18:10 UTC |
| addons | addons-924081 addons disable | addons-924081 | jenkins | v1.34.0 | 15 Sep 24 18:10 UTC | |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/15 17:57:04
Running on machine: ubuntu-20-agent
Binary: Built with gc go1.23.0 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0915 17:57:04.308539 19304 out.go:345] Setting OutFile to fd 1 ...
I0915 17:57:04.308786 19304 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0915 17:57:04.308795 19304 out.go:358] Setting ErrFile to fd 2...
I0915 17:57:04.308799 19304 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0915 17:57:04.308959 19304 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19648-11129/.minikube/bin
I0915 17:57:04.309553 19304 out.go:352] Setting JSON to false
I0915 17:57:04.310358 19304 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent","uptime":2370,"bootTime":1726420654,"procs":172,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1068-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0915 17:57:04.310443 19304 start.go:139] virtualization: kvm guest
I0915 17:57:04.312764 19304 out.go:177] * [addons-924081] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0915 17:57:04.314257 19304 notify.go:220] Checking for updates...
I0915 17:57:04.314269 19304 out.go:177] - MINIKUBE_LOCATION=19648
I0915 17:57:04.315814 19304 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0915 17:57:04.317230 19304 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19648-11129/kubeconfig
I0915 17:57:04.318630 19304 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19648-11129/.minikube
I0915 17:57:04.320105 19304 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0915 17:57:04.321479 19304 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0915 17:57:04.323273 19304 driver.go:394] Setting default libvirt URI to qemu:///system
I0915 17:57:04.346414 19304 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
I0915 17:57:04.346483 19304 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0915 17:57:04.399221 19304 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:30 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-15 17:57:04.390376351 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerError
s:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0915 17:57:04.399318 19304 docker.go:318] overlay module found
I0915 17:57:04.401441 19304 out.go:177] * Using the docker driver based on user configuration
I0915 17:57:04.402908 19304 start.go:297] selected driver: docker
I0915 17:57:04.402926 19304 start.go:901] validating driver "docker" against <nil>
I0915 17:57:04.402941 19304 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0915 17:57:04.403669 19304 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0915 17:57:04.450366 19304 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:30 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-15 17:57:04.441789997 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerError
s:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0915 17:57:04.450534 19304 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0915 17:57:04.450805 19304 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0915 17:57:04.452918 19304 out.go:177] * Using Docker driver with root privileges
I0915 17:57:04.454521 19304 cni.go:84] Creating CNI manager for ""
I0915 17:57:04.454582 19304 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0915 17:57:04.454593 19304 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0915 17:57:04.454660 19304 start.go:340] cluster config:
{Name:addons-924081 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-924081 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0915 17:57:04.456301 19304 out.go:177] * Starting "addons-924081" primary control-plane node in "addons-924081" cluster
I0915 17:57:04.457843 19304 cache.go:121] Beginning downloading kic base image for docker with docker
I0915 17:57:04.459411 19304 out.go:177] * Pulling base image v0.0.45-1726358845-19644 ...
I0915 17:57:04.460916 19304 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0915 17:57:04.460962 19304 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19648-11129/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4
I0915 17:57:04.460971 19304 cache.go:56] Caching tarball of preloaded images
I0915 17:57:04.461026 19304 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 in local docker daemon
I0915 17:57:04.461051 19304 preload.go:172] Found /home/jenkins/minikube-integration/19648-11129/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0915 17:57:04.461060 19304 cache.go:59] Finished verifying existence of preloaded tar for v1.31.1 on docker
I0915 17:57:04.461365 19304 profile.go:143] Saving config to /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/config.json ...
I0915 17:57:04.461388 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/config.json: {Name:mk891e370c5a5e03da92ab4a1a7be6c831238c28 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:04.478285 19304 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 to local cache
I0915 17:57:04.478422 19304 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 in local cache directory
I0915 17:57:04.478442 19304 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 in local cache directory, skipping pull
I0915 17:57:04.478452 19304 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 exists in cache, skipping pull
I0915 17:57:04.478461 19304 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 as a tarball
I0915 17:57:04.478466 19304 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 from local cache
I0915 17:57:16.550329 19304 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 from cached tarball
I0915 17:57:16.550362 19304 cache.go:194] Successfully downloaded all kic artifacts
I0915 17:57:16.550401 19304 start.go:360] acquireMachinesLock for addons-924081: {Name:mk882a2e9c8c2eb74b376399b00c4f80ae1e143d Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0915 17:57:16.550496 19304 start.go:364] duration metric: took 76.495µs to acquireMachinesLock for "addons-924081"
I0915 17:57:16.550532 19304 start.go:93] Provisioning new machine with config: &{Name:addons-924081 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-924081 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0915 17:57:16.550608 19304 start.go:125] createHost starting for "" (driver="docker")
I0915 17:57:16.553398 19304 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0915 17:57:16.553616 19304 start.go:159] libmachine.API.Create for "addons-924081" (driver="docker")
I0915 17:57:16.553648 19304 client.go:168] LocalClient.Create starting
I0915 17:57:16.553772 19304 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19648-11129/.minikube/certs/ca.pem
I0915 17:57:16.715619 19304 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19648-11129/.minikube/certs/cert.pem
I0915 17:57:16.804781 19304 cli_runner.go:164] Run: docker network inspect addons-924081 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0915 17:57:16.820095 19304 cli_runner.go:211] docker network inspect addons-924081 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0915 17:57:16.820177 19304 network_create.go:284] running [docker network inspect addons-924081] to gather additional debugging logs...
I0915 17:57:16.820200 19304 cli_runner.go:164] Run: docker network inspect addons-924081
W0915 17:57:16.835536 19304 cli_runner.go:211] docker network inspect addons-924081 returned with exit code 1
I0915 17:57:16.835564 19304 network_create.go:287] error running [docker network inspect addons-924081]: docker network inspect addons-924081: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-924081 not found
I0915 17:57:16.835590 19304 network_create.go:289] output of [docker network inspect addons-924081]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-924081 not found
** /stderr **
I0915 17:57:16.835720 19304 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0915 17:57:16.851386 19304 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001ad7520}
I0915 17:57:16.851428 19304 network_create.go:124] attempt to create docker network addons-924081 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0915 17:57:16.851476 19304 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-924081 addons-924081
I0915 17:57:16.910612 19304 network_create.go:108] docker network addons-924081 192.168.49.0/24 created
I0915 17:57:16.910644 19304 kic.go:121] calculated static IP "192.168.49.2" for the "addons-924081" container
I0915 17:57:16.910708 19304 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0915 17:57:16.925889 19304 cli_runner.go:164] Run: docker volume create addons-924081 --label name.minikube.sigs.k8s.io=addons-924081 --label created_by.minikube.sigs.k8s.io=true
I0915 17:57:16.942454 19304 oci.go:103] Successfully created a docker volume addons-924081
I0915 17:57:16.942529 19304 cli_runner.go:164] Run: docker run --rm --name addons-924081-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-924081 --entrypoint /usr/bin/test -v addons-924081:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -d /var/lib
I0915 17:57:21.740470 19304 cli_runner.go:217] Completed: docker run --rm --name addons-924081-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-924081 --entrypoint /usr/bin/test -v addons-924081:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -d /var/lib: (4.797906778s)
I0915 17:57:21.740493 19304 oci.go:107] Successfully prepared a docker volume addons-924081
I0915 17:57:21.740503 19304 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0915 17:57:21.740522 19304 kic.go:194] Starting extracting preloaded images to volume ...
I0915 17:57:21.740580 19304 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19648-11129/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-924081:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -I lz4 -xf /preloaded.tar -C /extractDir
I0915 17:57:25.642519 19304 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19648-11129/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-924081:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -I lz4 -xf /preloaded.tar -C /extractDir: (3.901891731s)
I0915 17:57:25.642551 19304 kic.go:203] duration metric: took 3.90202531s to extract preloaded images to volume ...
W0915 17:57:25.642712 19304 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0915 17:57:25.642848 19304 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0915 17:57:25.691199 19304 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-924081 --name addons-924081 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-924081 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-924081 --network addons-924081 --ip 192.168.49.2 --volume addons-924081:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0
I0915 17:57:25.989382 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Running}}
I0915 17:57:26.006661 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:26.023949 19304 cli_runner.go:164] Run: docker exec addons-924081 stat /var/lib/dpkg/alternatives/iptables
I0915 17:57:26.067786 19304 oci.go:144] the created container "addons-924081" has a running status.
I0915 17:57:26.067821 19304 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa...
I0915 17:57:26.216380 19304 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0915 17:57:26.238981 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:26.256666 19304 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0915 17:57:26.256688 19304 kic_runner.go:114] Args: [docker exec --privileged addons-924081 chown docker:docker /home/docker/.ssh/authorized_keys]
I0915 17:57:26.301516 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:26.318185 19304 machine.go:93] provisionDockerMachine start ...
I0915 17:57:26.318262 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:26.346937 19304 main.go:141] libmachine: Using SSH client type: native
I0915 17:57:26.347176 19304 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 17:57:26.347192 19304 main.go:141] libmachine: About to run SSH command:
hostname
I0915 17:57:26.347937 19304 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: read tcp 127.0.0.1:60170->127.0.0.1:32768: read: connection reset by peer
I0915 17:57:29.478002 19304 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-924081
I0915 17:57:29.478033 19304 ubuntu.go:169] provisioning hostname "addons-924081"
I0915 17:57:29.478094 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:29.494386 19304 main.go:141] libmachine: Using SSH client type: native
I0915 17:57:29.494551 19304 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 17:57:29.494564 19304 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-924081 && echo "addons-924081" | sudo tee /etc/hostname
I0915 17:57:29.637518 19304 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-924081
I0915 17:57:29.637591 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:29.655297 19304 main.go:141] libmachine: Using SSH client type: native
I0915 17:57:29.655457 19304 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 17:57:29.655473 19304 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-924081' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-924081/g' /etc/hosts;
else
echo '127.0.1.1 addons-924081' | sudo tee -a /etc/hosts;
fi
fi
I0915 17:57:29.786575 19304 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0915 17:57:29.786604 19304 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19648-11129/.minikube CaCertPath:/home/jenkins/minikube-integration/19648-11129/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19648-11129/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19648-11129/.minikube}
I0915 17:57:29.786639 19304 ubuntu.go:177] setting up certificates
I0915 17:57:29.786655 19304 provision.go:84] configureAuth start
I0915 17:57:29.786713 19304 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-924081
I0915 17:57:29.802003 19304 provision.go:143] copyHostCerts
I0915 17:57:29.802087 19304 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19648-11129/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19648-11129/.minikube/ca.pem (1082 bytes)
I0915 17:57:29.802207 19304 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19648-11129/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19648-11129/.minikube/cert.pem (1123 bytes)
I0915 17:57:29.802288 19304 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19648-11129/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19648-11129/.minikube/key.pem (1679 bytes)
I0915 17:57:29.802368 19304 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19648-11129/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19648-11129/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19648-11129/.minikube/certs/ca-key.pem org=jenkins.addons-924081 san=[127.0.0.1 192.168.49.2 addons-924081 localhost minikube]
I0915 17:57:29.955135 19304 provision.go:177] copyRemoteCerts
I0915 17:57:29.955208 19304 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0915 17:57:29.955252 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:29.971995 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:30.066949 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I0915 17:57:30.088131 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0915 17:57:30.109332 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0915 17:57:30.129931 19304 provision.go:87] duration metric: took 343.25986ms to configureAuth
I0915 17:57:30.129960 19304 ubuntu.go:193] setting minikube options for container-runtime
I0915 17:57:30.130102 19304 config.go:182] Loaded profile config "addons-924081": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0915 17:57:30.130145 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:30.146651 19304 main.go:141] libmachine: Using SSH client type: native
I0915 17:57:30.146839 19304 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 17:57:30.146851 19304 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0915 17:57:30.278969 19304 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0915 17:57:30.278991 19304 ubuntu.go:71] root file system type: overlay
I0915 17:57:30.279118 19304 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0915 17:57:30.279179 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:30.295356 19304 main.go:141] libmachine: Using SSH client type: native
I0915 17:57:30.295543 19304 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 17:57:30.295639 19304 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0915 17:57:30.436901 19304 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0915 17:57:30.436973 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:30.453652 19304 main.go:141] libmachine: Using SSH client type: native
I0915 17:57:30.453827 19304 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 17:57:30.453843 19304 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0915 17:57:31.135153 19304 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-09-06 12:06:41.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-15 17:57:30.431076713 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0915 17:57:31.135185 19304 machine.go:96] duration metric: took 4.8169796s to provisionDockerMachine
I0915 17:57:31.135200 19304 client.go:171] duration metric: took 14.581542314s to LocalClient.Create
I0915 17:57:31.135220 19304 start.go:167] duration metric: took 14.581604833s to libmachine.API.Create "addons-924081"
I0915 17:57:31.135231 19304 start.go:293] postStartSetup for "addons-924081" (driver="docker")
I0915 17:57:31.135240 19304 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0915 17:57:31.135292 19304 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0915 17:57:31.135333 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:31.151614 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:31.247237 19304 ssh_runner.go:195] Run: cat /etc/os-release
I0915 17:57:31.250046 19304 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0915 17:57:31.250084 19304 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0915 17:57:31.250094 19304 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0915 17:57:31.250103 19304 info.go:137] Remote host: Ubuntu 22.04.4 LTS
I0915 17:57:31.250113 19304 filesync.go:126] Scanning /home/jenkins/minikube-integration/19648-11129/.minikube/addons for local assets ...
I0915 17:57:31.250168 19304 filesync.go:126] Scanning /home/jenkins/minikube-integration/19648-11129/.minikube/files for local assets ...
I0915 17:57:31.250190 19304 start.go:296] duration metric: took 114.954264ms for postStartSetup
I0915 17:57:31.250461 19304 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-924081
I0915 17:57:31.266421 19304 profile.go:143] Saving config to /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/config.json ...
I0915 17:57:31.266878 19304 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0915 17:57:31.266934 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:31.283483 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:31.375214 19304 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0915 17:57:31.379065 19304 start.go:128] duration metric: took 14.828441455s to createHost
I0915 17:57:31.379091 19304 start.go:83] releasing machines lock for "addons-924081", held for 14.828581673s
I0915 17:57:31.379155 19304 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-924081
I0915 17:57:31.395144 19304 ssh_runner.go:195] Run: cat /version.json
I0915 17:57:31.395198 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:31.395261 19304 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0915 17:57:31.395312 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:31.412033 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:31.413936 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:31.572370 19304 ssh_runner.go:195] Run: systemctl --version
I0915 17:57:31.576279 19304 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0915 17:57:31.580034 19304 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0915 17:57:31.601338 19304 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0915 17:57:31.601400 19304 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0915 17:57:31.625793 19304 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0915 17:57:31.625820 19304 start.go:495] detecting cgroup driver to use...
I0915 17:57:31.625860 19304 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0915 17:57:31.625955 19304 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0915 17:57:31.640220 19304 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0915 17:57:31.648929 19304 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0915 17:57:31.657623 19304 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0915 17:57:31.657677 19304 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0915 17:57:31.666503 19304 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0915 17:57:31.675548 19304 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0915 17:57:31.684354 19304 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0915 17:57:31.692956 19304 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0915 17:57:31.701201 19304 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0915 17:57:31.710396 19304 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0915 17:57:31.719646 19304 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0915 17:57:31.728544 19304 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0915 17:57:31.735870 19304 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0915 17:57:31.743748 19304 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 17:57:31.819048 19304 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0915 17:57:31.911436 19304 start.go:495] detecting cgroup driver to use...
I0915 17:57:31.911490 19304 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0915 17:57:31.911537 19304 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0915 17:57:31.923378 19304 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0915 17:57:31.923468 19304 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0915 17:57:31.934300 19304 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0915 17:57:31.949595 19304 ssh_runner.go:195] Run: which cri-dockerd
I0915 17:57:31.952678 19304 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0915 17:57:31.960895 19304 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0915 17:57:31.977120 19304 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0915 17:57:32.060316 19304 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0915 17:57:32.159196 19304 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0915 17:57:32.159313 19304 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0915 17:57:32.174909 19304 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 17:57:32.252056 19304 ssh_runner.go:195] Run: sudo systemctl restart docker
I0915 17:57:32.491557 19304 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0915 17:57:32.501563 19304 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0915 17:57:32.511189 19304 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0915 17:57:32.594163 19304 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0915 17:57:32.674189 19304 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 17:57:32.746745 19304 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0915 17:57:32.758466 19304 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0915 17:57:32.767962 19304 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 17:57:32.840060 19304 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0915 17:57:32.899978 19304 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0915 17:57:32.900072 19304 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0915 17:57:32.903980 19304 start.go:563] Will wait 60s for crictl version
I0915 17:57:32.904038 19304 ssh_runner.go:195] Run: which crictl
I0915 17:57:32.907033 19304 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0915 17:57:32.938796 19304 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.2.1
RuntimeApiVersion: v1
I0915 17:57:32.938873 19304 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0915 17:57:32.961234 19304 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0915 17:57:32.986377 19304 out.go:235] * Preparing Kubernetes v1.31.1 on Docker 27.2.1 ...
I0915 17:57:32.986459 19304 cli_runner.go:164] Run: docker network inspect addons-924081 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0915 17:57:33.002562 19304 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0915 17:57:33.005908 19304 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0915 17:57:33.015547 19304 kubeadm.go:883] updating cluster {Name:addons-924081 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-924081 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0915 17:57:33.015680 19304 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0915 17:57:33.015736 19304 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0915 17:57:33.033509 19304 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0915 17:57:33.033531 19304 docker.go:615] Images already preloaded, skipping extraction
I0915 17:57:33.033602 19304 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0915 17:57:33.050641 19304 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0915 17:57:33.050663 19304 cache_images.go:84] Images are preloaded, skipping loading
I0915 17:57:33.050680 19304 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.1 docker true true} ...
I0915 17:57:33.050813 19304 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-924081 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.1 ClusterName:addons-924081 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0915 17:57:33.050876 19304 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0915 17:57:33.094941 19304 cni.go:84] Creating CNI manager for ""
I0915 17:57:33.094970 19304 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0915 17:57:33.094982 19304 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0915 17:57:33.095000 19304 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-924081 NodeName:addons-924081 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0915 17:57:33.095115 19304 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-924081"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0915 17:57:33.095167 19304 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.1
I0915 17:57:33.103187 19304 binaries.go:44] Found k8s binaries, skipping transfer
I0915 17:57:33.103248 19304 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0915 17:57:33.111010 19304 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0915 17:57:33.126679 19304 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0915 17:57:33.142034 19304 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0915 17:57:33.157513 19304 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0915 17:57:33.160572 19304 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0915 17:57:33.169962 19304 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 17:57:33.248922 19304 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0915 17:57:33.261323 19304 certs.go:68] Setting up /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081 for IP: 192.168.49.2
I0915 17:57:33.261343 19304 certs.go:194] generating shared ca certs ...
I0915 17:57:33.261358 19304 certs.go:226] acquiring lock for ca certs: {Name:mk64df53462b737f6bb192ffdd1f8219c712c8d9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:33.261482 19304 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19648-11129/.minikube/ca.key
I0915 17:57:33.828799 19304 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19648-11129/.minikube/ca.crt ...
I0915 17:57:33.828831 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/ca.crt: {Name:mkc86bdd0f64cf161ebb1f9d18f7e12f0930229b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:33.829021 19304 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19648-11129/.minikube/ca.key ...
I0915 17:57:33.829034 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/ca.key: {Name:mk71b60007278d168819bd274a4d8c0891031495 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:33.829135 19304 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19648-11129/.minikube/proxy-client-ca.key
I0915 17:57:33.968475 19304 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19648-11129/.minikube/proxy-client-ca.crt ...
I0915 17:57:33.968506 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/proxy-client-ca.crt: {Name:mk9add123944b044155f593aaca73a09db7f009c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:33.968694 19304 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19648-11129/.minikube/proxy-client-ca.key ...
I0915 17:57:33.968708 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/proxy-client-ca.key: {Name:mk3cd78008207692728a87ff13ea020c575f799f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:33.968806 19304 certs.go:256] generating profile certs ...
I0915 17:57:33.968877 19304 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/client.key
I0915 17:57:33.968895 19304 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/client.crt with IP's: []
I0915 17:57:34.138021 19304 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/client.crt ...
I0915 17:57:34.138053 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/client.crt: {Name:mkcf01dab9c3eab2bd9d85b6c5d4e5fd7f6abea5 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:34.138240 19304 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/client.key ...
I0915 17:57:34.138254 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/client.key: {Name:mkbce5b274e547b56ef0a170132afdc391aaa5e8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:34.138349 19304 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.key.67f312cf
I0915 17:57:34.138373 19304 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.crt.67f312cf with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0915 17:57:34.246594 19304 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.crt.67f312cf ...
I0915 17:57:34.246628 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.crt.67f312cf: {Name:mk3533e5a3b63037de3f1d7aa52276da7468cc89 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:34.246829 19304 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.key.67f312cf ...
I0915 17:57:34.246846 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.key.67f312cf: {Name:mk89d37c30c70ebc8f72c68d6cd8282f51cccd30 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:34.246959 19304 certs.go:381] copying /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.crt.67f312cf -> /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.crt
I0915 17:57:34.247072 19304 certs.go:385] copying /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.key.67f312cf -> /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.key
I0915 17:57:34.247146 19304 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/proxy-client.key
I0915 17:57:34.247175 19304 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/proxy-client.crt with IP's: []
I0915 17:57:34.376761 19304 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/proxy-client.crt ...
I0915 17:57:34.376801 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/proxy-client.crt: {Name:mke19914bd1151b8369b1920cd186fa0444a0ada Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:34.376979 19304 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/proxy-client.key ...
I0915 17:57:34.376993 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/proxy-client.key: {Name:mk653aea631ad864bf723125c711eed4ec5692fe Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:34.377470 19304 certs.go:484] found cert: /home/jenkins/minikube-integration/19648-11129/.minikube/certs/ca-key.pem (1679 bytes)
I0915 17:57:34.377526 19304 certs.go:484] found cert: /home/jenkins/minikube-integration/19648-11129/.minikube/certs/ca.pem (1082 bytes)
I0915 17:57:34.377563 19304 certs.go:484] found cert: /home/jenkins/minikube-integration/19648-11129/.minikube/certs/cert.pem (1123 bytes)
I0915 17:57:34.377600 19304 certs.go:484] found cert: /home/jenkins/minikube-integration/19648-11129/.minikube/certs/key.pem (1679 bytes)
I0915 17:57:34.378667 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0915 17:57:34.400410 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0915 17:57:34.420558 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0915 17:57:34.440643 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0915 17:57:34.460549 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0915 17:57:34.480824 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0915 17:57:34.501075 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0915 17:57:34.521324 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/profiles/addons-924081/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0915 17:57:34.541376 19304 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19648-11129/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0915 17:57:34.561745 19304 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0915 17:57:34.577223 19304 ssh_runner.go:195] Run: openssl version
I0915 17:57:34.582087 19304 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0915 17:57:34.590288 19304 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0915 17:57:34.593394 19304 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 15 17:57 /usr/share/ca-certificates/minikubeCA.pem
I0915 17:57:34.593437 19304 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0915 17:57:34.599626 19304 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0915 17:57:34.607833 19304 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0915 17:57:34.610719 19304 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0915 17:57:34.610775 19304 kubeadm.go:392] StartCluster: {Name:addons-924081 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-924081 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0915 17:57:34.610893 19304 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0915 17:57:34.627376 19304 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0915 17:57:34.635133 19304 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0915 17:57:34.642959 19304 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0915 17:57:34.643007 19304 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0915 17:57:34.650352 19304 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0915 17:57:34.650415 19304 kubeadm.go:157] found existing configuration files:
I0915 17:57:34.650458 19304 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0915 17:57:34.657662 19304 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0915 17:57:34.657718 19304 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0915 17:57:34.664809 19304 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0915 17:57:34.672029 19304 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0915 17:57:34.672080 19304 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0915 17:57:34.679113 19304 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0915 17:57:34.686221 19304 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0915 17:57:34.686267 19304 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0915 17:57:34.693329 19304 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0915 17:57:34.700813 19304 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0915 17:57:34.700862 19304 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0915 17:57:34.708301 19304 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0915 17:57:34.743204 19304 kubeadm.go:310] [init] Using Kubernetes version: v1.31.1
I0915 17:57:34.743257 19304 kubeadm.go:310] [preflight] Running pre-flight checks
I0915 17:57:34.763494 19304 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0915 17:57:34.763590 19304 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1068-gcp[0m
I0915 17:57:34.763635 19304 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0915 17:57:34.763713 19304 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0915 17:57:34.763804 19304 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0915 17:57:34.763891 19304 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0915 17:57:34.763966 19304 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0915 17:57:34.764032 19304 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0915 17:57:34.764108 19304 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0915 17:57:34.764177 19304 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0915 17:57:34.764246 19304 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0915 17:57:34.764315 19304 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0915 17:57:34.812995 19304 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0915 17:57:34.813129 19304 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0915 17:57:34.813241 19304 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0915 17:57:34.822555 19304 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0915 17:57:34.825581 19304 out.go:235] - Generating certificates and keys ...
I0915 17:57:34.825677 19304 kubeadm.go:310] [certs] Using existing ca certificate authority
I0915 17:57:34.825750 19304 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0915 17:57:35.001064 19304 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0915 17:57:35.497811 19304 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0915 17:57:35.693447 19304 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0915 17:57:35.922868 19304 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0915 17:57:36.090153 19304 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0915 17:57:36.090269 19304 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-924081 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0915 17:57:36.249285 19304 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0915 17:57:36.249430 19304 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-924081 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0915 17:57:36.427098 19304 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0915 17:57:36.489847 19304 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0915 17:57:36.687487 19304 kubeadm.go:310] [certs] Generating "sa" key and public key
I0915 17:57:36.687597 19304 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0915 17:57:36.874296 19304 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0915 17:57:36.999566 19304 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0915 17:57:37.309176 19304 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0915 17:57:37.386534 19304 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0915 17:57:37.685456 19304 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0915 17:57:37.685861 19304 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0915 17:57:37.688257 19304 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0915 17:57:37.690365 19304 out.go:235] - Booting up control plane ...
I0915 17:57:37.690454 19304 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0915 17:57:37.690523 19304 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0915 17:57:37.691054 19304 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0915 17:57:37.699869 19304 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0915 17:57:37.704977 19304 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0915 17:57:37.705036 19304 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0915 17:57:37.790653 19304 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0915 17:57:37.790781 19304 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0915 17:57:38.293028 19304 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 501.7377ms
I0915 17:57:38.293163 19304 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0915 17:57:42.794325 19304 kubeadm.go:310] [api-check] The API server is healthy after 4.501982833s
I0915 17:57:42.805490 19304 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0915 17:57:42.815042 19304 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0915 17:57:42.831132 19304 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0915 17:57:42.831399 19304 kubeadm.go:310] [mark-control-plane] Marking the node addons-924081 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0915 17:57:42.837754 19304 kubeadm.go:310] [bootstrap-token] Using token: frpd3s.5ayuoav6svg3tyrq
I0915 17:57:42.839041 19304 out.go:235] - Configuring RBAC rules ...
I0915 17:57:42.839186 19304 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0915 17:57:42.843325 19304 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0915 17:57:42.850055 19304 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0915 17:57:42.852413 19304 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0915 17:57:42.854608 19304 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0915 17:57:42.856994 19304 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0915 17:57:43.199889 19304 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0915 17:57:43.643324 19304 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0915 17:57:44.199722 19304 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0915 17:57:44.200527 19304 kubeadm.go:310]
I0915 17:57:44.200610 19304 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0915 17:57:44.200621 19304 kubeadm.go:310]
I0915 17:57:44.200696 19304 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0915 17:57:44.200704 19304 kubeadm.go:310]
I0915 17:57:44.200725 19304 kubeadm.go:310] mkdir -p $HOME/.kube
I0915 17:57:44.200781 19304 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0915 17:57:44.200825 19304 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0915 17:57:44.200843 19304 kubeadm.go:310]
I0915 17:57:44.200933 19304 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0915 17:57:44.200953 19304 kubeadm.go:310]
I0915 17:57:44.201025 19304 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0915 17:57:44.201037 19304 kubeadm.go:310]
I0915 17:57:44.201112 19304 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0915 17:57:44.201213 19304 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0915 17:57:44.201310 19304 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0915 17:57:44.201320 19304 kubeadm.go:310]
I0915 17:57:44.201446 19304 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0915 17:57:44.201541 19304 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0915 17:57:44.201550 19304 kubeadm.go:310]
I0915 17:57:44.201626 19304 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token frpd3s.5ayuoav6svg3tyrq \
I0915 17:57:44.201757 19304 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:8a792a950dea0de4c6def1e7426f15475332fd5b2be459c37ac9a68238375f24 \
I0915 17:57:44.201804 19304 kubeadm.go:310] --control-plane
I0915 17:57:44.201814 19304 kubeadm.go:310]
I0915 17:57:44.201930 19304 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0915 17:57:44.201939 19304 kubeadm.go:310]
I0915 17:57:44.202042 19304 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token frpd3s.5ayuoav6svg3tyrq \
I0915 17:57:44.202182 19304 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:8a792a950dea0de4c6def1e7426f15475332fd5b2be459c37ac9a68238375f24
I0915 17:57:44.204208 19304 kubeadm.go:310] W0915 17:57:34.740756 1924 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0915 17:57:44.204460 19304 kubeadm.go:310] W0915 17:57:34.741321 1924 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0915 17:57:44.204707 19304 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1068-gcp\n", err: exit status 1
I0915 17:57:44.204847 19304 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0915 17:57:44.204862 19304 cni.go:84] Creating CNI manager for ""
I0915 17:57:44.204883 19304 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0915 17:57:44.206671 19304 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0915 17:57:44.207801 19304 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0915 17:57:44.215668 19304 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0915 17:57:44.231561 19304 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0915 17:57:44.231656 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:44.231682 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-924081 minikube.k8s.io/updated_at=2024_09_15T17_57_44_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=6b3e75bb13951e1aa9da4105a14c95c8da7f2673 minikube.k8s.io/name=addons-924081 minikube.k8s.io/primary=true
I0915 17:57:44.296293 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:44.334968 19304 ops.go:34] apiserver oom_adj: -16
I0915 17:57:44.796522 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:45.297253 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:45.796957 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:46.296450 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:46.797158 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:47.297356 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:47.797014 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:48.297329 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:48.797243 19304 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 17:57:48.864971 19304 kubeadm.go:1113] duration metric: took 4.633375412s to wait for elevateKubeSystemPrivileges
I0915 17:57:48.865009 19304 kubeadm.go:394] duration metric: took 14.254237395s to StartCluster
I0915 17:57:48.865032 19304 settings.go:142] acquiring lock: {Name:mk8fb406764d83efd0c2a982185f31c6d8eb1dd9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:48.865143 19304 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19648-11129/kubeconfig
I0915 17:57:48.865572 19304 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19648-11129/kubeconfig: {Name:mk63f7e1b431103dccd36626008b13a19d1029e1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 17:57:48.865765 19304 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0915 17:57:48.865798 19304 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0915 17:57:48.865870 19304 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0915 17:57:48.865981 19304 addons.go:69] Setting yakd=true in profile "addons-924081"
I0915 17:57:48.865999 19304 addons.go:69] Setting default-storageclass=true in profile "addons-924081"
I0915 17:57:48.866023 19304 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-924081"
I0915 17:57:48.866031 19304 addons.go:69] Setting registry=true in profile "addons-924081"
I0915 17:57:48.866033 19304 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-924081"
I0915 17:57:48.866037 19304 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-924081"
I0915 17:57:48.866042 19304 addons.go:234] Setting addon registry=true in "addons-924081"
I0915 17:57:48.866035 19304 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-924081"
I0915 17:57:48.866049 19304 config.go:182] Loaded profile config "addons-924081": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0915 17:57:48.866050 19304 addons.go:69] Setting ingress-dns=true in profile "addons-924081"
I0915 17:57:48.866069 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.866077 19304 addons.go:69] Setting inspektor-gadget=true in profile "addons-924081"
I0915 17:57:48.866079 19304 addons.go:234] Setting addon ingress-dns=true in "addons-924081"
I0915 17:57:48.866088 19304 addons.go:234] Setting addon inspektor-gadget=true in "addons-924081"
I0915 17:57:48.866087 19304 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-924081"
I0915 17:57:48.866106 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.866118 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.866121 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.866130 19304 addons.go:69] Setting storage-provisioner=true in profile "addons-924081"
I0915 17:57:48.866148 19304 addons.go:234] Setting addon storage-provisioner=true in "addons-924081"
I0915 17:57:48.866171 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.866182 19304 addons.go:69] Setting volcano=true in profile "addons-924081"
I0915 17:57:48.866194 19304 addons.go:234] Setting addon volcano=true in "addons-924081"
I0915 17:57:48.866023 19304 addons.go:69] Setting helm-tiller=true in profile "addons-924081"
I0915 17:57:48.866215 19304 addons.go:234] Setting addon helm-tiller=true in "addons-924081"
I0915 17:57:48.866217 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.866239 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.866452 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.866596 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.866605 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.866069 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.866665 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.866665 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.866014 19304 addons.go:69] Setting gcp-auth=true in profile "addons-924081"
I0915 17:57:48.866678 19304 addons.go:69] Setting ingress=true in profile "addons-924081"
I0915 17:57:48.866690 19304 addons.go:234] Setting addon ingress=true in "addons-924081"
I0915 17:57:48.866693 19304 mustload.go:65] Loading cluster: addons-924081
I0915 17:57:48.866714 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.866903 19304 config.go:182] Loaded profile config "addons-924081": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0915 17:57:48.867055 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.867153 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.867230 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.866007 19304 addons.go:234] Setting addon yakd=true in "addons-924081"
I0915 17:57:48.867789 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.868305 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.866608 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.866005 19304 addons.go:69] Setting cloud-spanner=true in profile "addons-924081"
I0915 17:57:48.870040 19304 addons.go:234] Setting addon cloud-spanner=true in "addons-924081"
I0915 17:57:48.870076 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.870433 19304 addons.go:69] Setting volumesnapshots=true in profile "addons-924081"
I0915 17:57:48.870486 19304 addons.go:234] Setting addon volumesnapshots=true in "addons-924081"
I0915 17:57:48.870543 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.871158 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.866667 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.871972 19304 out.go:177] * Verifying Kubernetes components...
I0915 17:57:48.866120 19304 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-924081"
I0915 17:57:48.873738 19304 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-924081"
I0915 17:57:48.874412 19304 addons.go:69] Setting metrics-server=true in profile "addons-924081"
I0915 17:57:48.874436 19304 addons.go:234] Setting addon metrics-server=true in "addons-924081"
I0915 17:57:48.874470 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.875327 19304 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 17:57:48.866665 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.899982 19304 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0915 17:57:48.900626 19304 addons.go:234] Setting addon default-storageclass=true in "addons-924081"
I0915 17:57:48.900671 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.901164 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.901184 19304 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0915 17:57:48.901199 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0915 17:57:48.901247 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.903802 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.904247 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.904923 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.909944 19304 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0915 17:57:48.912296 19304 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0915 17:57:48.913932 19304 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0915 17:57:48.914801 19304 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0915 17:57:48.915446 19304 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0915 17:57:48.915471 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0915 17:57:48.915535 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.915924 19304 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0915 17:57:48.915941 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0915 17:57:48.915992 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.923012 19304 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0915 17:57:48.924483 19304 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0915 17:57:48.925760 19304 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0915 17:57:48.928039 19304 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0915 17:57:48.928073 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0915 17:57:48.928133 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.933168 19304 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0915 17:57:48.934870 19304 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0915 17:57:48.934892 19304 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0915 17:57:48.934957 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.935379 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.942599 19304 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0915 17:57:48.945877 19304 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0915 17:57:48.945901 19304 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0915 17:57:48.945972 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.953101 19304 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0915 17:57:48.956809 19304 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0915 17:57:48.956840 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0915 17:57:48.956900 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.967385 19304 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0915 17:57:48.967425 19304 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0915 17:57:48.967387 19304 out.go:177] - Using image docker.io/registry:2.8.3
I0915 17:57:48.967385 19304 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0915 17:57:48.969697 19304 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0915 17:57:48.969720 19304 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0915 17:57:48.969803 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.970585 19304 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0915 17:57:48.970702 19304 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0915 17:57:48.970713 19304 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0915 17:57:48.970794 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.971771 19304 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0915 17:57:48.973293 19304 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0915 17:57:48.973395 19304 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0915 17:57:48.973414 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0915 17:57:48.973464 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.975345 19304 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0915 17:57:48.976431 19304 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0915 17:57:48.977503 19304 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0915 17:57:48.979113 19304 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0915 17:57:48.979130 19304 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0915 17:57:48.979195 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.979293 19304 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-924081"
I0915 17:57:48.979329 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:48.979613 19304 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0915 17:57:48.980264 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:48.980958 19304 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0915 17:57:48.982109 19304 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0915 17:57:48.982299 19304 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0915 17:57:48.982313 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0915 17:57:48.982364 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.983195 19304 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0915 17:57:48.983214 19304 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0915 17:57:48.983342 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:48.996243 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:48.997602 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:48.998521 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.006098 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.007578 19304 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0915 17:57:49.009063 19304 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0915 17:57:49.009089 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0915 17:57:49.009149 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:49.011494 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.016020 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.017302 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.018172 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.020929 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.043599 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.044319 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.044565 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.046917 19304 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0915 17:57:49.048674 19304 out.go:177] - Using image docker.io/busybox:stable
I0915 17:57:49.049872 19304 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0915 17:57:49.049892 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0915 17:57:49.049949 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:49.050855 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
W0915 17:57:49.050875 19304 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0915 17:57:49.050904 19304 retry.go:31] will retry after 370.074531ms: ssh: handshake failed: EOF
W0915 17:57:49.051710 19304 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0915 17:57:49.051734 19304 retry.go:31] will retry after 186.847865ms: ssh: handshake failed: EOF
I0915 17:57:49.055480 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.070876 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:49.138742 19304 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0915 17:57:49.138816 19304 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
W0915 17:57:49.239783 19304 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0915 17:57:49.239814 19304 retry.go:31] will retry after 514.010771ms: ssh: handshake failed: EOF
I0915 17:57:49.335057 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0915 17:57:49.423838 19304 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0915 17:57:49.423863 19304 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0915 17:57:49.424206 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0915 17:57:49.426453 19304 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0915 17:57:49.426478 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0915 17:57:49.523896 19304 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0915 17:57:49.523979 19304 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0915 17:57:49.619586 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0915 17:57:49.643171 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0915 17:57:49.723332 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0915 17:57:49.728604 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0915 17:57:49.729014 19304 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0915 17:57:49.729036 19304 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0915 17:57:49.732445 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0915 17:57:49.737493 19304 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0915 17:57:49.737554 19304 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0915 17:57:49.737865 19304 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0915 17:57:49.737912 19304 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0915 17:57:49.738820 19304 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0915 17:57:49.738875 19304 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0915 17:57:49.824253 19304 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0915 17:57:49.824347 19304 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0915 17:57:49.832850 19304 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0915 17:57:49.832933 19304 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0915 17:57:50.022037 19304 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0915 17:57:50.022124 19304 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0915 17:57:50.039240 19304 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0915 17:57:50.039331 19304 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0915 17:57:50.130718 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0915 17:57:50.225053 19304 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0915 17:57:50.225141 19304 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0915 17:57:50.231657 19304 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0915 17:57:50.231744 19304 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0915 17:57:50.239021 19304 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0915 17:57:50.239099 19304 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0915 17:57:50.243614 19304 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0915 17:57:50.243677 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0915 17:57:50.521095 19304 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0915 17:57:50.521177 19304 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0915 17:57:50.526289 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0915 17:57:50.628948 19304 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0915 17:57:50.628979 19304 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0915 17:57:50.629296 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0915 17:57:50.635652 19304 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0915 17:57:50.635683 19304 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0915 17:57:50.934436 19304 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0915 17:57:50.934520 19304 ssh_runner.go:362] scp helm-tiller/helm-tiller-rbac.yaml --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0915 17:57:51.037727 19304 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0915 17:57:51.037761 19304 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0915 17:57:51.128095 19304 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0915 17:57:51.128175 19304 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0915 17:57:51.235190 19304 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (2.096338923s)
I0915 17:57:51.235275 19304 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0915 17:57:51.236505 19304 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (2.097717269s)
I0915 17:57:51.237601 19304 node_ready.go:35] waiting up to 6m0s for node "addons-924081" to be "Ready" ...
I0915 17:57:51.329027 19304 node_ready.go:49] node "addons-924081" has status "Ready":"True"
I0915 17:57:51.329114 19304 node_ready.go:38] duration metric: took 91.456442ms for node "addons-924081" to be "Ready" ...
I0915 17:57:51.329140 19304 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0915 17:57:51.345947 19304 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace to be "Ready" ...
I0915 17:57:51.520790 19304 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0915 17:57:51.520820 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0915 17:57:51.537777 19304 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0915 17:57:51.537863 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0915 17:57:51.636709 19304 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0915 17:57:51.636799 19304 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0915 17:57:51.740017 19304 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-924081" context rescaled to 1 replicas
I0915 17:57:51.824428 19304 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0915 17:57:51.824518 19304 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0915 17:57:51.835400 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0915 17:57:51.843445 19304 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0915 17:57:51.843532 19304 ssh_runner.go:362] scp helm-tiller/helm-tiller-svc.yaml --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0915 17:57:51.939297 19304 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0915 17:57:51.939382 19304 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0915 17:57:52.021511 19304 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0915 17:57:52.021608 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0915 17:57:52.137520 19304 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0915 17:57:52.137611 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0915 17:57:52.442595 19304 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0915 17:57:52.442677 19304 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0915 17:57:52.522952 19304 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0915 17:57:52.523042 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0915 17:57:52.622009 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0915 17:57:52.629205 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0915 17:57:52.924453 19304 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0915 17:57:52.924542 19304 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0915 17:57:52.926331 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (3.591189503s)
I0915 17:57:53.032394 19304 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0915 17:57:53.032423 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0915 17:57:53.321553 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0915 17:57:53.425220 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:57:53.521890 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0915 17:57:55.431497 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:57:56.023485 19304 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0915 17:57:56.023565 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:56.051965 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:56.927152 19304 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0915 17:57:57.122543 19304 addons.go:234] Setting addon gcp-auth=true in "addons-924081"
I0915 17:57:57.122654 19304 host.go:66] Checking if "addons-924081" exists ...
I0915 17:57:57.123205 19304 cli_runner.go:164] Run: docker container inspect addons-924081 --format={{.State.Status}}
I0915 17:57:57.152996 19304 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0915 17:57:57.153040 19304 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-924081
I0915 17:57:57.168993 19304 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19648-11129/.minikube/machines/addons-924081/id_rsa Username:docker}
I0915 17:57:57.435030 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:57:58.333391 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (8.90914838s)
I0915 17:57:58.333431 19304 addons.go:475] Verifying addon ingress=true in "addons-924081"
I0915 17:57:58.334026 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (8.690824448s)
I0915 17:57:58.334122 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (8.610760371s)
I0915 17:57:58.334574 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (8.714307894s)
I0915 17:57:58.336175 19304 out.go:177] * Verifying ingress addon...
I0915 17:57:58.339644 19304 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0915 17:57:58.346312 19304 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0915 17:57:58.346330 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:57:58.927229 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:57:59.347543 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:57:59.850926 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:57:59.928610 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:00.345775 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:00.845038 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:01.433087 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:01.441534 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.712856931s)
I0915 17:58:01.441647 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (11.70912328s)
I0915 17:58:01.441715 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (11.310962423s)
I0915 17:58:01.441876 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (10.812551205s)
I0915 17:58:01.441899 19304 addons.go:475] Verifying addon registry=true in "addons-924081"
I0915 17:58:01.441937 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (9.606424958s)
I0915 17:58:01.442018 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (10.915448408s)
I0915 17:58:01.442040 19304 addons.go:475] Verifying addon metrics-server=true in "addons-924081"
I0915 17:58:01.442097 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (8.820047775s)
I0915 17:58:01.442227 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (8.812932385s)
W0915 17:58:01.442276 19304 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0915 17:58:01.442304 19304 retry.go:31] will retry after 352.53232ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0915 17:58:01.521275 19304 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-924081 service yakd-dashboard -n yakd-dashboard
I0915 17:58:01.521393 19304 out.go:177] * Verifying registry addon...
I0915 17:58:01.525374 19304 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0915 17:58:01.528994 19304 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0915 17:58:01.529018 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:01.795447 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0915 17:58:01.922482 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:02.043257 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:02.349812 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:02.430039 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:02.530733 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:02.540466 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (9.218762454s)
I0915 17:58:02.540546 19304 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-924081"
I0915 17:58:02.540819 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (9.018887808s)
I0915 17:58:02.540873 19304 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.38785262s)
I0915 17:58:02.542420 19304 out.go:177] * Verifying csi-hostpath-driver addon...
I0915 17:58:02.542424 19304 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0915 17:58:02.544319 19304 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0915 17:58:02.545678 19304 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0915 17:58:02.547011 19304 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0915 17:58:02.547038 19304 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0915 17:58:02.631803 19304 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0915 17:58:02.631894 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:02.648707 19304 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0915 17:58:02.648791 19304 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0915 17:58:02.755620 19304 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0915 17:58:02.755648 19304 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0915 17:58:02.837137 19304 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0915 17:58:02.844630 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:03.030295 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:03.050794 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:03.344008 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:03.529892 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:03.550241 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:03.843607 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:04.029461 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:04.131316 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:04.331231 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.535731057s)
I0915 17:58:04.347743 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:04.445050 19304 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.607863736s)
I0915 17:58:04.447158 19304 addons.go:475] Verifying addon gcp-auth=true in "addons-924081"
I0915 17:58:04.448901 19304 out.go:177] * Verifying gcp-auth addon...
I0915 17:58:04.451469 19304 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0915 17:58:04.453862 19304 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0915 17:58:04.529576 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:04.550285 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:04.844290 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:04.852073 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:05.028651 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:05.050048 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:05.344078 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:05.555800 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:05.556652 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:05.845212 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:06.029253 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:06.051076 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:06.344091 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:06.529394 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:06.549877 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:06.844147 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:07.028919 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:07.052592 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:07.343387 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:07.351660 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:07.556879 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:07.557759 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:07.844228 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:08.029170 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:08.048910 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:08.343676 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:08.529817 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:08.550247 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:08.843867 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:09.028392 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:09.049318 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:09.343021 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:09.528168 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:09.550104 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:09.843773 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:09.850621 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:10.029177 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:10.050842 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:10.343803 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:10.528987 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:10.549982 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:10.843825 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:11.028817 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:11.050376 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:11.343766 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:11.556726 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:11.557062 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:11.843830 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:11.851305 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:12.029827 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:12.050448 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:12.343815 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:12.529478 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:12.549765 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:12.843620 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:13.029553 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:13.049277 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:13.344252 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:13.528468 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:13.549570 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:13.843756 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:13.851929 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:14.028943 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:14.050363 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:14.344079 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:14.529058 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:14.550859 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:14.843955 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:15.029611 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:15.050167 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:15.343981 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:15.528577 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:15.550635 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:15.890542 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:15.901327 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:16.028709 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:16.049657 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:16.343375 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:16.528622 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:16.550342 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:16.844062 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:17.028543 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:17.049543 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:17.343010 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:17.528744 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:17.549675 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:17.843282 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:18.029065 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:18.050107 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:18.343749 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:18.351222 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:18.529314 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:18.549779 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:18.843610 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:19.028574 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:19.050168 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:19.345370 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:19.529692 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:19.550527 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:19.844585 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:20.029403 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:20.050235 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:20.344542 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:20.351499 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:20.528739 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:20.549952 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:20.843236 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:21.028606 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:21.049660 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:21.343877 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:21.528734 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:21.550476 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:21.844019 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:22.056138 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:22.056764 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:22.344254 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:22.528970 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:22.550459 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:22.844293 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:22.851305 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:23.055759 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:23.056368 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:23.343527 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:23.529297 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:23.550283 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:23.843890 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:24.028282 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:24.049316 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:24.344365 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:24.529085 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:24.550572 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:24.844131 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:24.851793 19304 pod_ready.go:103] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"False"
I0915 17:58:25.029023 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:25.050261 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:25.344389 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:25.528782 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:25.549415 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:25.842971 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:26.029006 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:26.049767 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:26.343456 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:26.350879 19304 pod_ready.go:93] pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace has status "Ready":"True"
I0915 17:58:26.350897 19304 pod_ready.go:82] duration metric: took 35.004917128s for pod "coredns-7c65d6cfc9-8r5p2" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.350908 19304 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-nrsn5" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.352279 19304 pod_ready.go:98] error getting pod "coredns-7c65d6cfc9-nrsn5" in "kube-system" namespace (skipping!): pods "coredns-7c65d6cfc9-nrsn5" not found
I0915 17:58:26.352296 19304 pod_ready.go:82] duration metric: took 1.381723ms for pod "coredns-7c65d6cfc9-nrsn5" in "kube-system" namespace to be "Ready" ...
E0915 17:58:26.352304 19304 pod_ready.go:67] WaitExtra: waitPodCondition: error getting pod "coredns-7c65d6cfc9-nrsn5" in "kube-system" namespace (skipping!): pods "coredns-7c65d6cfc9-nrsn5" not found
I0915 17:58:26.352311 19304 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-924081" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.355863 19304 pod_ready.go:93] pod "etcd-addons-924081" in "kube-system" namespace has status "Ready":"True"
I0915 17:58:26.355878 19304 pod_ready.go:82] duration metric: took 3.56196ms for pod "etcd-addons-924081" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.355886 19304 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-924081" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.359162 19304 pod_ready.go:93] pod "kube-apiserver-addons-924081" in "kube-system" namespace has status "Ready":"True"
I0915 17:58:26.359178 19304 pod_ready.go:82] duration metric: took 3.286396ms for pod "kube-apiserver-addons-924081" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.359189 19304 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-924081" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.362373 19304 pod_ready.go:93] pod "kube-controller-manager-addons-924081" in "kube-system" namespace has status "Ready":"True"
I0915 17:58:26.362388 19304 pod_ready.go:82] duration metric: took 3.192079ms for pod "kube-controller-manager-addons-924081" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.362399 19304 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-djh4b" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.528663 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:26.549737 19304 pod_ready.go:93] pod "kube-proxy-djh4b" in "kube-system" namespace has status "Ready":"True"
I0915 17:58:26.549763 19304 pod_ready.go:82] duration metric: took 187.35433ms for pod "kube-proxy-djh4b" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.549775 19304 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-924081" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.549920 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:26.843646 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:26.950311 19304 pod_ready.go:93] pod "kube-scheduler-addons-924081" in "kube-system" namespace has status "Ready":"True"
I0915 17:58:26.950333 19304 pod_ready.go:82] duration metric: took 400.50753ms for pod "kube-scheduler-addons-924081" in "kube-system" namespace to be "Ready" ...
I0915 17:58:26.950342 19304 pod_ready.go:79] waiting up to 6m0s for pod "nvidia-device-plugin-daemonset-nhvqc" in "kube-system" namespace to be "Ready" ...
I0915 17:58:27.029166 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:27.049046 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:27.342901 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:27.349261 19304 pod_ready.go:93] pod "nvidia-device-plugin-daemonset-nhvqc" in "kube-system" namespace has status "Ready":"True"
I0915 17:58:27.349283 19304 pod_ready.go:82] duration metric: took 398.933608ms for pod "nvidia-device-plugin-daemonset-nhvqc" in "kube-system" namespace to be "Ready" ...
I0915 17:58:27.349294 19304 pod_ready.go:39] duration metric: took 36.020106565s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0915 17:58:27.349348 19304 api_server.go:52] waiting for apiserver process to appear ...
I0915 17:58:27.349415 19304 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0915 17:58:27.362636 19304 api_server.go:72] duration metric: took 38.496810723s to wait for apiserver process to appear ...
I0915 17:58:27.362664 19304 api_server.go:88] waiting for apiserver healthz status ...
I0915 17:58:27.362681 19304 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0915 17:58:27.366156 19304 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0915 17:58:27.367018 19304 api_server.go:141] control plane version: v1.31.1
I0915 17:58:27.367039 19304 api_server.go:131] duration metric: took 4.370242ms to wait for apiserver health ...
I0915 17:58:27.367046 19304 system_pods.go:43] waiting for kube-system pods to appear ...
I0915 17:58:27.529146 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:27.550009 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:27.556476 19304 system_pods.go:59] 18 kube-system pods found
I0915 17:58:27.556504 19304 system_pods.go:61] "coredns-7c65d6cfc9-8r5p2" [1956ba7b-4677-4f65-9946-3ecb8a5db57b] Running
I0915 17:58:27.556512 19304 system_pods.go:61] "csi-hostpath-attacher-0" [d9e70213-5a19-4ff0-a34c-aa58ece1ff75] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0915 17:58:27.556519 19304 system_pods.go:61] "csi-hostpath-resizer-0" [4abcdbf1-e874-41c7-9d35-a1745bb07c16] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0915 17:58:27.556526 19304 system_pods.go:61] "csi-hostpathplugin-86wk4" [4cb1592c-4b1d-42bc-969c-43e9937ec9b8] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0915 17:58:27.556532 19304 system_pods.go:61] "etcd-addons-924081" [083f6778-402e-4aa4-a4db-3e8873271227] Running
I0915 17:58:27.556537 19304 system_pods.go:61] "kube-apiserver-addons-924081" [620085ab-d25f-4483-8ef3-31b69a03c7db] Running
I0915 17:58:27.556543 19304 system_pods.go:61] "kube-controller-manager-addons-924081" [370a7ed1-3f52-41c3-96e7-3cece8bc350a] Running
I0915 17:58:27.556549 19304 system_pods.go:61] "kube-ingress-dns-minikube" [661ba882-e028-4cdd-bb37-8ee95de61c69] Running
I0915 17:58:27.556554 19304 system_pods.go:61] "kube-proxy-djh4b" [06adfc98-36ba-4500-b6a0-2887a8b024b3] Running
I0915 17:58:27.556559 19304 system_pods.go:61] "kube-scheduler-addons-924081" [d8141836-e57b-4f9c-91d3-c8228b01a81e] Running
I0915 17:58:27.556564 19304 system_pods.go:61] "metrics-server-84c5f94fbc-g29nd" [d0a4650f-3b55-4081-b127-353cca2c9570] Running
I0915 17:58:27.556572 19304 system_pods.go:61] "nvidia-device-plugin-daemonset-nhvqc" [47a2d060-ff2d-4161-9188-f26d8cb11aa1] Running
I0915 17:58:27.556576 19304 system_pods.go:61] "registry-66c9cd494c-85p89" [7ddd4a6c-0bb9-4cdd-b2c2-6a358cc36131] Running
I0915 17:58:27.556584 19304 system_pods.go:61] "registry-proxy-lrwnn" [727ad348-b4a0-40a9-a423-cac288b38182] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0915 17:58:27.556595 19304 system_pods.go:61] "snapshot-controller-56fcc65765-kr6b5" [1ceebd9b-a2ce-49df-8214-2a12a64b390e] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0915 17:58:27.556605 19304 system_pods.go:61] "snapshot-controller-56fcc65765-zrwhx" [aee77165-1ded-4628-8591-f325e4697bbf] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0915 17:58:27.556612 19304 system_pods.go:61] "storage-provisioner" [b92320a2-149a-4ffe-9494-baa89c63524d] Running
I0915 17:58:27.556617 19304 system_pods.go:61] "tiller-deploy-b48cc5f79-8kwdn" [44490bdf-edf8-403c-a16b-77e4a27b2aca] Pending / Ready:ContainersNotReady (containers with unready status: [tiller]) / ContainersReady:ContainersNotReady (containers with unready status: [tiller])
I0915 17:58:27.556625 19304 system_pods.go:74] duration metric: took 189.574073ms to wait for pod list to return data ...
I0915 17:58:27.556633 19304 default_sa.go:34] waiting for default service account to be created ...
I0915 17:58:27.749146 19304 default_sa.go:45] found service account: "default"
I0915 17:58:27.749168 19304 default_sa.go:55] duration metric: took 192.529995ms for default service account to be created ...
I0915 17:58:27.749177 19304 system_pods.go:116] waiting for k8s-apps to be running ...
I0915 17:58:27.843251 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:27.956624 19304 system_pods.go:86] 18 kube-system pods found
I0915 17:58:27.956652 19304 system_pods.go:89] "coredns-7c65d6cfc9-8r5p2" [1956ba7b-4677-4f65-9946-3ecb8a5db57b] Running
I0915 17:58:27.956664 19304 system_pods.go:89] "csi-hostpath-attacher-0" [d9e70213-5a19-4ff0-a34c-aa58ece1ff75] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0915 17:58:27.956672 19304 system_pods.go:89] "csi-hostpath-resizer-0" [4abcdbf1-e874-41c7-9d35-a1745bb07c16] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0915 17:58:27.956683 19304 system_pods.go:89] "csi-hostpathplugin-86wk4" [4cb1592c-4b1d-42bc-969c-43e9937ec9b8] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0915 17:58:27.956690 19304 system_pods.go:89] "etcd-addons-924081" [083f6778-402e-4aa4-a4db-3e8873271227] Running
I0915 17:58:27.956699 19304 system_pods.go:89] "kube-apiserver-addons-924081" [620085ab-d25f-4483-8ef3-31b69a03c7db] Running
I0915 17:58:27.956708 19304 system_pods.go:89] "kube-controller-manager-addons-924081" [370a7ed1-3f52-41c3-96e7-3cece8bc350a] Running
I0915 17:58:27.956719 19304 system_pods.go:89] "kube-ingress-dns-minikube" [661ba882-e028-4cdd-bb37-8ee95de61c69] Running
I0915 17:58:27.956724 19304 system_pods.go:89] "kube-proxy-djh4b" [06adfc98-36ba-4500-b6a0-2887a8b024b3] Running
I0915 17:58:27.956733 19304 system_pods.go:89] "kube-scheduler-addons-924081" [d8141836-e57b-4f9c-91d3-c8228b01a81e] Running
I0915 17:58:27.956739 19304 system_pods.go:89] "metrics-server-84c5f94fbc-g29nd" [d0a4650f-3b55-4081-b127-353cca2c9570] Running
I0915 17:58:27.956745 19304 system_pods.go:89] "nvidia-device-plugin-daemonset-nhvqc" [47a2d060-ff2d-4161-9188-f26d8cb11aa1] Running
I0915 17:58:27.956750 19304 system_pods.go:89] "registry-66c9cd494c-85p89" [7ddd4a6c-0bb9-4cdd-b2c2-6a358cc36131] Running
I0915 17:58:27.956759 19304 system_pods.go:89] "registry-proxy-lrwnn" [727ad348-b4a0-40a9-a423-cac288b38182] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0915 17:58:27.956765 19304 system_pods.go:89] "snapshot-controller-56fcc65765-kr6b5" [1ceebd9b-a2ce-49df-8214-2a12a64b390e] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0915 17:58:27.956778 19304 system_pods.go:89] "snapshot-controller-56fcc65765-zrwhx" [aee77165-1ded-4628-8591-f325e4697bbf] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0915 17:58:27.956784 19304 system_pods.go:89] "storage-provisioner" [b92320a2-149a-4ffe-9494-baa89c63524d] Running
I0915 17:58:27.956795 19304 system_pods.go:89] "tiller-deploy-b48cc5f79-8kwdn" [44490bdf-edf8-403c-a16b-77e4a27b2aca] Pending / Ready:ContainersNotReady (containers with unready status: [tiller]) / ContainersReady:ContainersNotReady (containers with unready status: [tiller])
I0915 17:58:27.956808 19304 system_pods.go:126] duration metric: took 207.62584ms to wait for k8s-apps to be running ...
I0915 17:58:27.956818 19304 system_svc.go:44] waiting for kubelet service to be running ....
I0915 17:58:27.956867 19304 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0915 17:58:27.971680 19304 system_svc.go:56] duration metric: took 14.853041ms WaitForService to wait for kubelet
I0915 17:58:27.971712 19304 kubeadm.go:582] duration metric: took 39.105889227s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0915 17:58:27.971736 19304 node_conditions.go:102] verifying NodePressure condition ...
I0915 17:58:28.028992 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:28.050327 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:28.150240 19304 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0915 17:58:28.150275 19304 node_conditions.go:123] node cpu capacity is 8
I0915 17:58:28.150292 19304 node_conditions.go:105] duration metric: took 178.550174ms to run NodePressure ...
I0915 17:58:28.150306 19304 start.go:241] waiting for startup goroutines ...
I0915 17:58:28.344166 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:28.529010 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:28.550336 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:28.843787 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:29.029638 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:29.050210 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:29.344539 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:29.529581 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:29.550277 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:29.844452 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:30.032032 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:30.049850 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:30.343431 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:30.529707 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:30.549480 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:30.843095 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:31.055511 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 17:58:31.056032 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:31.343735 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:31.529003 19304 kapi.go:107] duration metric: took 30.003631074s to wait for kubernetes.io/minikube-addons=registry ...
I0915 17:58:31.550085 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:31.844241 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:32.049958 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:32.344342 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:32.549220 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:32.844891 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:33.050116 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:33.344323 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:33.550324 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:33.844337 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:34.050204 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:34.343414 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:34.556590 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:34.843643 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:35.049921 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:35.343858 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:35.549615 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:35.844518 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:36.050359 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:36.343893 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:36.550268 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:36.844655 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:37.049956 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:37.343540 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:37.550657 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:37.844885 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:38.049400 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:38.343861 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:38.549688 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:38.843350 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:39.050517 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:39.344598 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:39.549791 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:39.843996 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:40.050218 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:40.344034 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:40.550641 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:40.843588 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:41.056010 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:41.343848 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:41.550510 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:41.844034 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:42.050530 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:42.362714 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:42.623978 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:42.844113 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:43.049863 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:43.342906 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:43.549853 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:43.843982 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:44.050347 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:44.343788 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:44.550666 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:44.844949 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:45.050206 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:45.343586 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:45.549947 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:45.843875 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:46.052222 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:46.343926 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:46.550212 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:46.843942 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:47.050092 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:47.343649 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:47.608262 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:47.843971 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:48.049987 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:48.343494 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:48.551040 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:48.843718 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:49.050628 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:49.343762 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:49.550896 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:49.843467 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:50.114434 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:50.344135 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:50.551442 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:50.844666 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:51.050832 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:51.343415 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:51.550185 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:51.844604 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:52.050637 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:52.344158 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:52.549497 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:52.845006 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:53.057238 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:53.344502 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:53.550080 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:53.844458 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:54.050074 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:54.343296 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:54.556998 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:54.844732 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:55.050405 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:55.343887 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:55.567009 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:55.843933 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:56.050506 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:56.344468 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:56.550055 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:56.844250 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:57.057673 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:57.360365 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:57.550041 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:57.844656 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:58.050306 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:58.344484 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:58.557370 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:58.844609 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:59.050615 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 17:58:59.343389 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:58:59.550156 19304 kapi.go:107] duration metric: took 57.004478741s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0915 17:58:59.843691 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:00.342865 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:00.843087 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:01.343410 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:01.844106 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:02.343477 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:02.844488 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:03.344031 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:03.843891 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:04.344403 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:04.844708 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:05.345487 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:05.844082 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:06.343583 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:06.844551 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:07.343415 19304 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 17:59:07.845794 19304 kapi.go:107] duration metric: took 1m9.506150381s to wait for app.kubernetes.io/name=ingress-nginx ...
I0915 17:59:27.954409 19304 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0915 17:59:27.954433 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:28.454426 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:28.954535 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:29.454742 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:29.954551 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:30.454348 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:30.955265 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:31.455499 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:31.954455 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:32.454102 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:32.955689 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:33.454789 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:33.955063 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:34.455187 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:34.954447 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:35.454532 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:35.954422 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:36.454288 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:36.955508 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:37.454526 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:37.954607 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:38.454384 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:38.955643 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:39.454934 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:39.954423 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:40.454014 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:40.954838 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:41.455095 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:41.954934 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:42.454436 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:42.954270 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:43.455288 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:43.955377 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:44.454218 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:44.955023 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:45.455057 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:45.954929 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:46.454690 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:46.955001 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:47.454914 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:47.955181 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:48.454965 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:48.954450 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:49.454792 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:49.954579 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:50.454214 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:50.955554 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:51.454856 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:51.954816 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:52.454502 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:52.954343 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:53.454547 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:53.956702 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:54.454951 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:54.955279 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:55.454350 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:55.955196 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:56.454952 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:56.954777 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:57.455206 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:57.954804 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:58.454490 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:58.954371 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:59.454673 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 17:59:59.954476 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:00.454140 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:00.955103 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:01.454846 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:01.954437 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:02.454174 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:02.955408 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:03.454710 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:03.954486 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:04.454604 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:04.954628 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:05.454771 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:05.954464 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:06.454135 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:06.954654 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:07.454828 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:07.956250 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:08.455022 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:08.955239 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:09.456112 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:09.954646 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:10.454781 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:10.954736 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:11.454599 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:11.954969 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:12.454646 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:12.954837 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:13.454936 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:13.954503 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:14.454655 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:14.954999 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:15.455247 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:15.955099 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:16.455480 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:16.955115 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:17.454919 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:17.954912 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:18.454442 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:18.954385 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:19.454903 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:19.955098 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:20.454725 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:20.954627 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:21.454576 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:21.954507 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:22.454304 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:22.955220 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:23.455451 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:23.954851 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:24.454941 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:24.955173 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:25.455174 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:25.954956 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:26.454954 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:26.955286 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:27.455796 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:27.953991 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:28.454472 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:28.955298 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:29.454480 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:29.953986 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:30.454737 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:30.954655 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:31.454702 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:31.954718 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:32.454633 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:32.955861 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:33.455192 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:33.954702 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:34.455050 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:34.954490 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:35.454527 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:35.954492 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:36.454381 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:36.955138 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:37.455721 19304 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 18:00:37.955002 19304 kapi.go:107] duration metric: took 2m33.503531422s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0915 18:00:37.956862 19304 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-924081 cluster.
I0915 18:00:37.958378 19304 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0915 18:00:37.959887 19304 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0915 18:00:37.961345 19304 out.go:177] * Enabled addons: storage-provisioner, nvidia-device-plugin, ingress-dns, storage-provisioner-rancher, volcano, cloud-spanner, metrics-server, helm-tiller, yakd, default-storageclass, inspektor-gadget, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0915 18:00:37.962788 19304 addons.go:510] duration metric: took 2m49.096897282s for enable addons: enabled=[storage-provisioner nvidia-device-plugin ingress-dns storage-provisioner-rancher volcano cloud-spanner metrics-server helm-tiller yakd default-storageclass inspektor-gadget volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0915 18:00:37.962840 19304 start.go:246] waiting for cluster config update ...
I0915 18:00:37.962870 19304 start.go:255] writing updated cluster config ...
I0915 18:00:37.963178 19304 ssh_runner.go:195] Run: rm -f paused
I0915 18:00:38.013286 19304 start.go:600] kubectl: 1.31.0, cluster: 1.31.1 (minor skew: 0)
I0915 18:00:38.015166 19304 out.go:177] * Done! kubectl is now configured to use "addons-924081" cluster and "default" namespace by default
==> Docker <==
Sep 15 18:09:59 addons-924081 dockerd[1342]: time="2024-09-15T18:09:59.645231356Z" level=info msg="ignoring event" container=6d0fadb9deebfd14b70831ef8db3b6ad56c37cd194f986e5ea2c86eb06b14836 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:09:59 addons-924081 dockerd[1342]: time="2024-09-15T18:09:59.829302639Z" level=info msg="ignoring event" container=652ef040ff28a424626a4bbcb186d80e79d14cc75da789f95bb0231734d997e9 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:09:59 addons-924081 dockerd[1342]: time="2024-09-15T18:09:59.839170418Z" level=info msg="ignoring event" container=bf4bc0c5191c154f4502856c6f49c47cedc23e7335274a36881735ec8002aa60 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:00 addons-924081 dockerd[1342]: time="2024-09-15T18:10:00.034128742Z" level=info msg="ignoring event" container=68fbf1b0779f8f23541458bad09664c3b3d553a264ae628c77bdff20e7dfad08 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:00 addons-924081 dockerd[1342]: time="2024-09-15T18:10:00.635822841Z" level=info msg="ignoring event" container=debbf53b4ab3326b68cdb1d6a063a27edaa92afb5e669466b7659f1dca714a70 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:00 addons-924081 dockerd[1342]: time="2024-09-15T18:10:00.649568484Z" level=info msg="ignoring event" container=4f9ed26676baea430f41a447a86c0c60b5521a4b29e2f5b87b26743fca58989f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:00 addons-924081 dockerd[1342]: time="2024-09-15T18:10:00.818651884Z" level=info msg="ignoring event" container=9dda5677134430853c93af4fc0ece6b20af4f3ea92c1ae937eca2d8f300045a7 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:00 addons-924081 dockerd[1342]: time="2024-09-15T18:10:00.861345895Z" level=info msg="ignoring event" container=99f851b499b502039ae4b1c74084fde375e9bae7e21c30021818afbacf86ba31 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:01 addons-924081 cri-dockerd[1607]: time="2024-09-15T18:10:01Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/fc929ba652317e6d6c550ee97115acc2fd87e73ba7ddd9b866e7e8133d6d894d/resolv.conf as [nameserver 10.96.0.10 search headlamp.svc.cluster.local svc.cluster.local cluster.local europe-west1-b.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 15 18:10:02 addons-924081 dockerd[1342]: time="2024-09-15T18:10:02.173527531Z" level=warning msg="reference for unknown type: " digest="sha256:899d106eeb55b0afc4ee6e51c03bc4418de0bd0e79c39744d4d0d751aae6a971" remote="ghcr.io/headlamp-k8s/headlamp@sha256:899d106eeb55b0afc4ee6e51c03bc4418de0bd0e79c39744d4d0d751aae6a971"
Sep 15 18:10:05 addons-924081 dockerd[1342]: time="2024-09-15T18:10:05.384044956Z" level=info msg="ignoring event" container=4159ab2ccb2e3aeb85d1eac68c5e98f6e1a429bebef74f96b99c23d3cf5b2f5f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:11 addons-924081 cri-dockerd[1607]: time="2024-09-15T18:10:11Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/efce7c5474bd6ec9e23d209fd820f6a0caa9fc6105e69c9893911aaf50336e58/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local europe-west1-b.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 15 18:10:12 addons-924081 cri-dockerd[1607]: time="2024-09-15T18:10:12Z" level=info msg="Pulling image ghcr.io/headlamp-k8s/headlamp:v0.25.0@sha256:899d106eeb55b0afc4ee6e51c03bc4418de0bd0e79c39744d4d0d751aae6a971: c9712956fa89: Downloading [============================> ] 19.54MB/33.82MB"
Sep 15 18:10:20 addons-924081 cri-dockerd[1607]: time="2024-09-15T18:10:20Z" level=info msg="Stop pulling image ghcr.io/headlamp-k8s/headlamp:v0.25.0@sha256:899d106eeb55b0afc4ee6e51c03bc4418de0bd0e79c39744d4d0d751aae6a971: Status: Downloaded newer image for ghcr.io/headlamp-k8s/headlamp@sha256:899d106eeb55b0afc4ee6e51c03bc4418de0bd0e79c39744d4d0d751aae6a971"
Sep 15 18:10:20 addons-924081 dockerd[1342]: time="2024-09-15T18:10:20.337656414Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 15 18:10:20 addons-924081 dockerd[1342]: time="2024-09-15T18:10:20.409416814Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 15 18:10:23 addons-924081 cri-dockerd[1607]: time="2024-09-15T18:10:23Z" level=info msg="Stop pulling image docker.io/nginx:alpine: Status: Downloaded newer image for nginx:alpine"
Sep 15 18:10:30 addons-924081 dockerd[1342]: time="2024-09-15T18:10:30.416178833Z" level=info msg="ignoring event" container=694ac69cd8f251d649349c8414fae36813e5a1524b509d0550e0cd7cbd773f9b module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:30 addons-924081 cri-dockerd[1607]: time="2024-09-15T18:10:30Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/f0bae8633a79b0cafaab23533e340bb9eb4ea231a7c7a6db3b6d2a7252dcd628/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local europe-west1-b.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 15 18:10:31 addons-924081 dockerd[1342]: time="2024-09-15T18:10:31.035912607Z" level=info msg="ignoring event" container=99c979201ef97c6c87baf1c857f890b602203a87ec0726983cbb938e98f2ce6c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:31 addons-924081 dockerd[1342]: time="2024-09-15T18:10:31.131956765Z" level=info msg="ignoring event" container=86acdab2983506e5442611ad9ce6fdc9c6ee456f465bb913b0fed4e6390161f0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:31 addons-924081 dockerd[1342]: time="2024-09-15T18:10:31.189807192Z" level=info msg="ignoring event" container=71985211be19f5926e075487d7f21b805b3edd6f00b346d8e924adee636fc916 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:31 addons-924081 dockerd[1342]: time="2024-09-15T18:10:31.323282482Z" level=info msg="ignoring event" container=466f137f35e637dbae2f34b6d117ccb4d09869177a3fcd59e37ac80bee303505 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:31 addons-924081 dockerd[1342]: time="2024-09-15T18:10:31.451131912Z" level=info msg="ignoring event" container=6d1e4de06db8f2f6fabcfdf8fb7586505ba8cc6f134c275294b36f4e8ad49fac module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 18:10:31 addons-924081 dockerd[1342]: time="2024-09-15T18:10:31.500116116Z" level=info msg="ignoring event" container=eb4793a100e16552c35fed357b2d2fe67ce964e70632c742091ad567711c4399 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
ef93d4aa88788 nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf 9 seconds ago Running nginx 0 efce7c5474bd6 nginx
52d9e76e8846c ghcr.io/headlamp-k8s/headlamp@sha256:899d106eeb55b0afc4ee6e51c03bc4418de0bd0e79c39744d4d0d751aae6a971 12 seconds ago Running headlamp 0 fc929ba652317 headlamp-57fb76fcdb-pcmm5
05408f6b2b993 gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 ba961def1a2bd gcp-auth-89d5ffd79-d9sgx
aaa57b3f73755 registry.k8s.io/ingress-nginx/controller@sha256:d5f8217feeac4887cb1ed21f27c2674e58be06bd8f5184cacea2a69abaf78dce 11 minutes ago Running controller 0 3a9ef92e1c7c3 ingress-nginx-controller-bc57996ff-vr8pl
ed27aad05ae80 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited patch 0 593a0a466fabb ingress-nginx-admission-patch-qdf6l
a13bca916718f registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 c44a704d322bb ingress-nginx-admission-create-7pvj5
182d5cd1be62d 6e38f40d628db 12 minutes ago Running storage-provisioner 0 e9d0f91b1ef6d storage-provisioner
f3fdc5a30a919 c69fa2e9cbf5f 12 minutes ago Running coredns 0 67261b45a9146 coredns-7c65d6cfc9-8r5p2
305fd1c940590 60c005f310ff3 12 minutes ago Running kube-proxy 0 d43465abd9705 kube-proxy-djh4b
66f12aeb790eb 6bab7719df100 12 minutes ago Running kube-apiserver 0 77fe25ea23941 kube-apiserver-addons-924081
9f19c328253e6 2e96e5913fc06 12 minutes ago Running etcd 0 4f2c562feff0f etcd-addons-924081
ee0638f0200f9 175ffd71cce3d 12 minutes ago Running kube-controller-manager 0 1607b21ed569e kube-controller-manager-addons-924081
511083959df7e 9aa1fad941575 12 minutes ago Running kube-scheduler 0 38555d6b4a477 kube-scheduler-addons-924081
==> controller_ingress [aaa57b3f7375] <==
I0915 17:59:08.665830 7 controller.go:224] "Initial sync, sleeping for 1 second"
I0915 17:59:08.665921 7 event.go:377] Event(v1.ObjectReference{Kind:"Pod", Namespace:"ingress-nginx", Name:"ingress-nginx-controller-bc57996ff-vr8pl", UID:"ecfcd2ea-e4ba-4096-ae5f-eb89b0cd9dd4", APIVersion:"v1", ResourceVersion:"1289", FieldPath:""}): type: 'Normal' reason: 'RELOAD' NGINX reload triggered due to a change in configuration
W0915 18:10:10.777065 7 controller.go:1110] Error obtaining Endpoints for Service "default/nginx": no object matching key "default/nginx" in local store
I0915 18:10:10.795254 7 admission.go:149] processed ingress via admission controller {testedIngressLength:1 testedIngressTime:0.018s renderingIngressLength:1 renderingIngressTime:0.001s admissionTime:0.019s testedConfigurationSize:18.1kB}
I0915 18:10:10.795290 7 main.go:107] "successfully validated configuration, accepting" ingress="default/nginx-ingress"
I0915 18:10:10.798473 7 store.go:440] "Found valid IngressClass" ingress="default/nginx-ingress" ingressclass="nginx"
I0915 18:10:10.798649 7 event.go:377] Event(v1.ObjectReference{Kind:"Ingress", Namespace:"default", Name:"nginx-ingress", UID:"6dee2574-d6b1-4bc2-8652-d1656c7b5f07", APIVersion:"networking.k8s.io/v1", ResourceVersion:"2974", FieldPath:""}): type: 'Normal' reason: 'Sync' Scheduled for sync
W0915 18:10:10.798867 7 controller.go:1110] Error obtaining Endpoints for Service "default/nginx": no object matching key "default/nginx" in local store
I0915 18:10:10.798948 7 controller.go:193] "Configuration changes detected, backend reload required"
I0915 18:10:10.836201 7 controller.go:213] "Backend successfully reloaded"
I0915 18:10:10.836429 7 event.go:377] Event(v1.ObjectReference{Kind:"Pod", Namespace:"ingress-nginx", Name:"ingress-nginx-controller-bc57996ff-vr8pl", UID:"ecfcd2ea-e4ba-4096-ae5f-eb89b0cd9dd4", APIVersion:"v1", ResourceVersion:"1289", FieldPath:""}): type: 'Normal' reason: 'RELOAD' NGINX reload triggered due to a change in configuration
W0915 18:10:14.133663 7 controller.go:1216] Service "default/nginx" does not have any active Endpoint.
I0915 18:10:14.133807 7 controller.go:193] "Configuration changes detected, backend reload required"
I0915 18:10:14.170097 7 controller.go:213] "Backend successfully reloaded"
I0915 18:10:14.170289 7 event.go:377] Event(v1.ObjectReference{Kind:"Pod", Namespace:"ingress-nginx", Name:"ingress-nginx-controller-bc57996ff-vr8pl", UID:"ecfcd2ea-e4ba-4096-ae5f-eb89b0cd9dd4", APIVersion:"v1", ResourceVersion:"1289", FieldPath:""}): type: 'Normal' reason: 'RELOAD' NGINX reload triggered due to a change in configuration
W0915 18:10:21.530018 7 controller.go:1216] Service "default/nginx" does not have any active Endpoint.
W0915 18:10:30.442466 7 controller.go:1110] Error obtaining Endpoints for Service "kube-system/hello-world-app": no object matching key "kube-system/hello-world-app" in local store
I0915 18:10:30.468168 7 admission.go:149] processed ingress via admission controller {testedIngressLength:2 testedIngressTime:0.026s renderingIngressLength:2 renderingIngressTime:0s admissionTime:0.026s testedConfigurationSize:26.2kB}
I0915 18:10:30.468196 7 main.go:107] "successfully validated configuration, accepting" ingress="kube-system/example-ingress"
I0915 18:10:30.526709 7 store.go:440] "Found valid IngressClass" ingress="kube-system/example-ingress" ingressclass="nginx"
I0915 18:10:30.526902 7 event.go:377] Event(v1.ObjectReference{Kind:"Ingress", Namespace:"kube-system", Name:"example-ingress", UID:"1644cac0-6956-422a-85ab-ef81a36f33ad", APIVersion:"networking.k8s.io/v1", ResourceVersion:"3044", FieldPath:""}): type: 'Normal' reason: 'Sync' Scheduled for sync
I0915 18:10:31.530220 7 controller.go:193] "Configuration changes detected, backend reload required"
I0915 18:10:31.578672 7 controller.go:213] "Backend successfully reloaded"
I0915 18:10:31.579035 7 event.go:377] Event(v1.ObjectReference{Kind:"Pod", Namespace:"ingress-nginx", Name:"ingress-nginx-controller-bc57996ff-vr8pl", UID:"ecfcd2ea-e4ba-4096-ae5f-eb89b0cd9dd4", APIVersion:"v1", ResourceVersion:"1289", FieldPath:""}): type: 'Normal' reason: 'RELOAD' NGINX reload triggered due to a change in configuration
10.244.0.1 - - [15/Sep/2024:18:10:30 +0000] "GET / HTTP/1.1" 200 615 "-" "curl/7.81.0" 81 0.000 [default-nginx-80] [] 10.244.0.37:80 615 0.001 200 8dbbfb156245d59455be6a1f8a0514aa
==> coredns [f3fdc5a30a91] <==
[INFO] 10.244.0.22:40155 - 54502 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.005100416s
[INFO] 10.244.0.22:55349 - 53091 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005365613s
[INFO] 10.244.0.22:40155 - 41295 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005048283s
[INFO] 10.244.0.22:57499 - 23314 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005445218s
[INFO] 10.244.0.22:42673 - 62895 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005362245s
[INFO] 10.244.0.22:46216 - 37926 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005473042s
[INFO] 10.244.0.22:50828 - 46710 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003571893s
[INFO] 10.244.0.22:37729 - 24951 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005408336s
[INFO] 10.244.0.22:50132 - 5192 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00569195s
[INFO] 10.244.0.22:57499 - 55638 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003046684s
[INFO] 10.244.0.22:37729 - 37944 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.00404348s
[INFO] 10.244.0.22:46216 - 3305 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004440124s
[INFO] 10.244.0.22:42673 - 12631 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004371989s
[INFO] 10.244.0.22:57499 - 47270 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000206402s
[INFO] 10.244.0.22:37729 - 39026 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000056709s
[INFO] 10.244.0.22:55349 - 28210 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004843783s
[INFO] 10.244.0.22:40155 - 21984 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004346055s
[INFO] 10.244.0.22:50828 - 42891 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005009271s
[INFO] 10.244.0.22:50132 - 9731 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004254036s
[INFO] 10.244.0.22:50132 - 5871 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000078617s
[INFO] 10.244.0.22:46216 - 39960 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000054386s
[INFO] 10.244.0.22:55349 - 20118 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000044347s
[INFO] 10.244.0.22:42673 - 28304 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000041256s
[INFO] 10.244.0.22:40155 - 2135 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000042175s
[INFO] 10.244.0.22:50828 - 49741 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000041548s
==> describe nodes <==
Name: addons-924081
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-924081
kubernetes.io/os=linux
minikube.k8s.io/commit=6b3e75bb13951e1aa9da4105a14c95c8da7f2673
minikube.k8s.io/name=addons-924081
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_15T17_57_44_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-924081
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Sun, 15 Sep 2024 17:57:41 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-924081
AcquireTime: <unset>
RenewTime: Sun, 15 Sep 2024 18:10:29 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Sun, 15 Sep 2024 18:10:18 +0000 Sun, 15 Sep 2024 17:57:39 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Sun, 15 Sep 2024 18:10:18 +0000 Sun, 15 Sep 2024 17:57:39 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Sun, 15 Sep 2024 18:10:18 +0000 Sun, 15 Sep 2024 17:57:39 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Sun, 15 Sep 2024 18:10:18 +0000 Sun, 15 Sep 2024 17:57:41 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-924081
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
System Info:
Machine ID: eccf7df5bf9a4dc784472dbbf26344c8
System UUID: 7e2b3237-3df5-42a6-b42c-1a16f34283e0
Boot ID: c04e1fd2-9f8e-4626-99c0-4aa7783c27aa
Kernel Version: 5.15.0-1068-gcp
OS Image: Ubuntu 22.04.4 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.2.1
Kubelet Version: v1.31.1
Kube-Proxy Version: v1.31.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (13 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m16s
default hello-world-app-55bf9c44b4-xz47d 0 (0%) 0 (0%) 0 (0%) 0 (0%) 2s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 22s
gcp-auth gcp-auth-89d5ffd79-d9sgx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
headlamp headlamp-57fb76fcdb-pcmm5 0 (0%) 0 (0%) 0 (0%) 0 (0%) 31s
ingress-nginx ingress-nginx-controller-bc57996ff-vr8pl 100m (1%) 0 (0%) 90Mi (0%) 0 (0%) 12m
kube-system coredns-7c65d6cfc9-8r5p2 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-924081 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-924081 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-924081 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-djh4b 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-924081 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 850m (10%) 0 (0%)
memory 260Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeHasSufficientMemory 12m (x8 over 12m) kubelet Node addons-924081 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m (x7 over 12m) kubelet Node addons-924081 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m (x7 over 12m) kubelet Node addons-924081 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-924081 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-924081 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-924081 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-924081 event: Registered Node addons-924081 in Controller
==> dmesg <==
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 22 56 0b f4 44 87 08 06
[ +2.411845] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff ae 9b 9c 20 66 90 08 06
[ +6.110541] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff f2 dc f0 c5 4d 48 08 06
[ +0.070706] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff d2 75 de c9 64 42 08 06
[ +0.220930] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 76 23 7e 80 66 f1 08 06
[Sep15 17:59] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 5e 98 84 8b 95 48 08 06
[Sep15 18:00] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 5a 85 a8 ef 88 84 08 06
[ +0.105563] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff de ab c0 e4 84 7d 08 06
[ +29.129381] IPv4: martian source 10.244.0.1 from 10.244.0.26, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff b2 53 e4 fe cd a8 08 06
[ +0.000499] IPv4: martian source 10.244.0.26 from 10.244.0.3, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 82 2c 6b 33 6e 7c 08 06
[Sep15 18:09] IPv4: martian source 10.244.0.1 from 10.244.0.35, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff 3a 34 d0 4c bb 8f 08 06
[Sep15 18:10] IPv4: martian source 10.244.0.37 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 5e 98 84 8b 95 48 08 06
[ +1.372552] IPv4: martian source 10.244.0.22 from 10.244.0.3, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 82 2c 6b 33 6e 7c 08 06
==> etcd [9f19c328253e] <==
{"level":"info","ts":"2024-09-15T17:57:40.057823Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became pre-candidate at term 1"}
{"level":"info","ts":"2024-09-15T17:57:40.057841Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgPreVoteResp from aec36adc501070cc at term 1"}
{"level":"info","ts":"2024-09-15T17:57:40.057862Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became candidate at term 2"}
{"level":"info","ts":"2024-09-15T17:57:40.057869Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-15T17:57:40.057878Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-09-15T17:57:40.057886Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-15T17:57:40.058742Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-15T17:57:40.059466Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-15T17:57:40.059470Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-924081 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-09-15T17:57:40.059490Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-15T17:57:40.059807Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-15T17:57:40.059888Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-15T17:57:40.060048Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-15T17:57:40.060126Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-15T17:57:40.060154Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-15T17:57:40.060727Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-15T17:57:40.060802Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-15T17:57:40.061592Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-09-15T17:57:40.061991Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"warn","ts":"2024-09-15T17:58:15.888207Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"111.122676ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/events/gcp-auth/gcp-auth-89d5ffd79.17f57ca5f76c3fd9\" ","response":"range_response_count:1 size:928"}
{"level":"info","ts":"2024-09-15T17:58:15.888319Z","caller":"traceutil/trace.go:171","msg":"trace[418854937] range","detail":"{range_begin:/registry/events/gcp-auth/gcp-auth-89d5ffd79.17f57ca5f76c3fd9; range_end:; response_count:1; response_revision:1028; }","duration":"111.272227ms","start":"2024-09-15T17:58:15.777033Z","end":"2024-09-15T17:58:15.888305Z","steps":["trace[418854937] 'range keys from in-memory index tree' (duration: 110.997919ms)"],"step_count":1}
{"level":"info","ts":"2024-09-15T18:07:40.143589Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1913}
{"level":"info","ts":"2024-09-15T18:07:40.168196Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1913,"took":"24.089331ms","hash":3826931549,"current-db-size-bytes":8896512,"current-db-size":"8.9 MB","current-db-size-in-use-bytes":5033984,"current-db-size-in-use":"5.0 MB"}
{"level":"info","ts":"2024-09-15T18:07:40.168252Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":3826931549,"revision":1913,"compact-revision":-1}
{"level":"info","ts":"2024-09-15T18:10:20.609387Z","caller":"traceutil/trace.go:171","msg":"trace[1839691361] transaction","detail":"{read_only:false; response_revision:3001; number_of_response:1; }","duration":"118.418774ms","start":"2024-09-15T18:10:20.490944Z","end":"2024-09-15T18:10:20.609362Z","steps":["trace[1839691361] 'process raft request' (duration: 60.597159ms)","trace[1839691361] 'compare' (duration: 57.69822ms)"],"step_count":2}
==> gcp-auth [05408f6b2b99] <==
2024/09/15 18:01:16 Ready to write response ...
2024/09/15 18:09:19 Ready to marshal response ...
2024/09/15 18:09:19 Ready to write response ...
2024/09/15 18:09:19 Ready to marshal response ...
2024/09/15 18:09:19 Ready to write response ...
2024/09/15 18:09:28 Ready to marshal response ...
2024/09/15 18:09:28 Ready to write response ...
2024/09/15 18:09:29 Ready to marshal response ...
2024/09/15 18:09:29 Ready to write response ...
2024/09/15 18:09:30 Ready to marshal response ...
2024/09/15 18:09:30 Ready to write response ...
2024/09/15 18:09:43 Ready to marshal response ...
2024/09/15 18:09:43 Ready to write response ...
2024/09/15 18:09:54 Ready to marshal response ...
2024/09/15 18:09:54 Ready to write response ...
2024/09/15 18:10:01 Ready to marshal response ...
2024/09/15 18:10:01 Ready to write response ...
2024/09/15 18:10:01 Ready to marshal response ...
2024/09/15 18:10:01 Ready to write response ...
2024/09/15 18:10:01 Ready to marshal response ...
2024/09/15 18:10:01 Ready to write response ...
2024/09/15 18:10:10 Ready to marshal response ...
2024/09/15 18:10:10 Ready to write response ...
2024/09/15 18:10:30 Ready to marshal response ...
2024/09/15 18:10:30 Ready to write response ...
==> kernel <==
18:10:32 up 52 min, 0 users, load average: 0.84, 0.43, 0.30
Linux addons-924081 5.15.0-1068-gcp #76~20.04.1-Ubuntu SMP Tue Aug 20 15:52:45 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.4 LTS"
==> kube-apiserver [66f12aeb790e] <==
E0915 18:09:29.992520 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
E0915 18:09:29.999594 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
E0915 18:09:30.006284 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
I0915 18:09:37.598366 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
E0915 18:09:45.006230 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
I0915 18:10:00.388021 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0915 18:10:00.388079 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0915 18:10:00.401084 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0915 18:10:00.401138 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0915 18:10:00.402240 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0915 18:10:00.402286 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0915 18:10:00.422938 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0915 18:10:00.422987 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0915 18:10:00.523086 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0915 18:10:00.523161 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0915 18:10:01.418985 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
I0915 18:10:01.481701 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.96.199.10"}
W0915 18:10:01.523187 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
W0915 18:10:01.539463 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
I0915 18:10:05.340066 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0915 18:10:06.458166 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0915 18:10:10.796054 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0915 18:10:11.020774 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.96.39.176"}
I0915 18:10:25.238514 1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
I0915 18:10:30.619628 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.110.38.43"}
==> kube-controller-manager [ee0638f0200f] <==
I0915 18:10:18.517932 1 shared_informer.go:313] Waiting for caches to sync for resource quota
I0915 18:10:18.517975 1 shared_informer.go:320] Caches are synced for resource quota
W0915 18:10:18.631560 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 18:10:18.631604 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0915 18:10:18.823827 1 shared_informer.go:313] Waiting for caches to sync for garbage collector
I0915 18:10:18.823867 1 shared_informer.go:320] Caches are synced for garbage collector
W0915 18:10:20.318884 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 18:10:20.318934 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0915 18:10:21.135442 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 18:10:21.135482 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0915 18:10:21.525323 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="headlamp/headlamp-57fb76fcdb" duration="58.728µs"
I0915 18:10:21.541288 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="headlamp/headlamp-57fb76fcdb" duration="5.796058ms"
I0915 18:10:21.541413 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="headlamp/headlamp-57fb76fcdb" duration="81.122µs"
W0915 18:10:23.747598 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 18:10:23.747648 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0915 18:10:25.808777 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 18:10:25.808822 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0915 18:10:30.446271 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="15.556326ms"
I0915 18:10:30.452816 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="6.493954ms"
I0915 18:10:30.452908 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="50.182µs"
I0915 18:10:30.455547 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="44.841µs"
I0915 18:10:30.950436 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="3.724µs"
I0915 18:10:32.460138 1 job_controller.go:568] "enqueueing job" logger="job-controller" key="ingress-nginx/ingress-nginx-admission-create" delay="0s"
I0915 18:10:32.462107 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="ingress-nginx/ingress-nginx-controller-bc57996ff" duration="5.015µs"
I0915 18:10:32.464544 1 job_controller.go:568] "enqueueing job" logger="job-controller" key="ingress-nginx/ingress-nginx-admission-patch" delay="0s"
==> kube-proxy [305fd1c94059] <==
I0915 17:57:51.731198 1 server_linux.go:66] "Using iptables proxy"
I0915 17:57:52.237960 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0915 17:57:52.238037 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0915 17:57:52.727695 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0915 17:57:52.727764 1 server_linux.go:169] "Using iptables Proxier"
I0915 17:57:52.732410 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0915 17:57:52.732966 1 server.go:483] "Version info" version="v1.31.1"
I0915 17:57:52.732984 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0915 17:57:52.734509 1 config.go:199] "Starting service config controller"
I0915 17:57:52.734544 1 shared_informer.go:313] Waiting for caches to sync for service config
I0915 17:57:52.734587 1 config.go:105] "Starting endpoint slice config controller"
I0915 17:57:52.734594 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0915 17:57:52.734610 1 config.go:328] "Starting node config controller"
I0915 17:57:52.734624 1 shared_informer.go:313] Waiting for caches to sync for node config
I0915 17:57:52.835003 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0915 17:57:52.835050 1 shared_informer.go:320] Caches are synced for service config
I0915 17:57:52.835293 1 shared_informer.go:320] Caches are synced for node config
==> kube-scheduler [511083959df7] <==
E0915 17:57:41.243906 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
E0915 17:57:41.243906 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0915 17:57:41.244133 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
W0915 17:57:41.244168 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
W0915 17:57:41.244175 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0915 17:57:41.244192 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0915 17:57:41.244157 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
E0915 17:57:41.244199 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0915 17:57:41.244215 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0915 17:57:41.244260 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0915 17:57:41.244281 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
W0915 17:57:41.244338 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0915 17:57:41.244354 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0915 17:57:41.244358 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0915 17:57:41.244307 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0915 17:57:41.244416 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0915 17:57:41.244311 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0915 17:57:41.244446 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0915 17:57:42.085652 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0915 17:57:42.085693 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0915 17:57:42.227113 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0915 17:57:42.227153 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0915 17:57:42.384190 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0915 17:57:42.384231 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
I0915 17:57:42.842161 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 15 18:10:30 addons-924081 kubelet[2451]: I0915 18:10:30.636254 2451 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e86fc63e-1e75-477a-89d2-6c1f90366215-kube-api-access-fwfdd" (OuterVolumeSpecName: "kube-api-access-fwfdd") pod "e86fc63e-1e75-477a-89d2-6c1f90366215" (UID: "e86fc63e-1e75-477a-89d2-6c1f90366215"). InnerVolumeSpecName "kube-api-access-fwfdd". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 15 18:10:30 addons-924081 kubelet[2451]: I0915 18:10:30.731571 2451 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-fwfdd\" (UniqueName: \"kubernetes.io/projected/e86fc63e-1e75-477a-89d2-6c1f90366215-kube-api-access-fwfdd\") on node \"addons-924081\" DevicePath \"\""
Sep 15 18:10:30 addons-924081 kubelet[2451]: I0915 18:10:30.731615 2451 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/e86fc63e-1e75-477a-89d2-6c1f90366215-gcp-creds\") on node \"addons-924081\" DevicePath \"\""
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.437988 2451 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhkj2\" (UniqueName: \"kubernetes.io/projected/7ddd4a6c-0bb9-4cdd-b2c2-6a358cc36131-kube-api-access-bhkj2\") pod \"7ddd4a6c-0bb9-4cdd-b2c2-6a358cc36131\" (UID: \"7ddd4a6c-0bb9-4cdd-b2c2-6a358cc36131\") "
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.439976 2451 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ddd4a6c-0bb9-4cdd-b2c2-6a358cc36131-kube-api-access-bhkj2" (OuterVolumeSpecName: "kube-api-access-bhkj2") pod "7ddd4a6c-0bb9-4cdd-b2c2-6a358cc36131" (UID: "7ddd4a6c-0bb9-4cdd-b2c2-6a358cc36131"). InnerVolumeSpecName "kube-api-access-bhkj2". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.537707 2451 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e86fc63e-1e75-477a-89d2-6c1f90366215" path="/var/lib/kubelet/pods/e86fc63e-1e75-477a-89d2-6c1f90366215/volumes"
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.538236 2451 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnh8h\" (UniqueName: \"kubernetes.io/projected/727ad348-b4a0-40a9-a423-cac288b38182-kube-api-access-hnh8h\") pod \"727ad348-b4a0-40a9-a423-cac288b38182\" (UID: \"727ad348-b4a0-40a9-a423-cac288b38182\") "
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.538315 2451 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-bhkj2\" (UniqueName: \"kubernetes.io/projected/7ddd4a6c-0bb9-4cdd-b2c2-6a358cc36131-kube-api-access-bhkj2\") on node \"addons-924081\" DevicePath \"\""
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.541852 2451 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/727ad348-b4a0-40a9-a423-cac288b38182-kube-api-access-hnh8h" (OuterVolumeSpecName: "kube-api-access-hnh8h") pod "727ad348-b4a0-40a9-a423-cac288b38182" (UID: "727ad348-b4a0-40a9-a423-cac288b38182"). InnerVolumeSpecName "kube-api-access-hnh8h". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.639792 2451 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgmpr\" (UniqueName: \"kubernetes.io/projected/661ba882-e028-4cdd-bb37-8ee95de61c69-kube-api-access-jgmpr\") pod \"661ba882-e028-4cdd-bb37-8ee95de61c69\" (UID: \"661ba882-e028-4cdd-bb37-8ee95de61c69\") "
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.639914 2451 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-hnh8h\" (UniqueName: \"kubernetes.io/projected/727ad348-b4a0-40a9-a423-cac288b38182-kube-api-access-hnh8h\") on node \"addons-924081\" DevicePath \"\""
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.644645 2451 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/661ba882-e028-4cdd-bb37-8ee95de61c69-kube-api-access-jgmpr" (OuterVolumeSpecName: "kube-api-access-jgmpr") pod "661ba882-e028-4cdd-bb37-8ee95de61c69" (UID: "661ba882-e028-4cdd-bb37-8ee95de61c69"). InnerVolumeSpecName "kube-api-access-jgmpr". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.728798 2451 scope.go:117] "RemoveContainer" containerID="86acdab2983506e5442611ad9ce6fdc9c6ee456f465bb913b0fed4e6390161f0"
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.740648 2451 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-jgmpr\" (UniqueName: \"kubernetes.io/projected/661ba882-e028-4cdd-bb37-8ee95de61c69-kube-api-access-jgmpr\") on node \"addons-924081\" DevicePath \"\""
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.757398 2451 scope.go:117] "RemoveContainer" containerID="86acdab2983506e5442611ad9ce6fdc9c6ee456f465bb913b0fed4e6390161f0"
Sep 15 18:10:31 addons-924081 kubelet[2451]: E0915 18:10:31.822058 2451 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 86acdab2983506e5442611ad9ce6fdc9c6ee456f465bb913b0fed4e6390161f0" containerID="86acdab2983506e5442611ad9ce6fdc9c6ee456f465bb913b0fed4e6390161f0"
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.822106 2451 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"86acdab2983506e5442611ad9ce6fdc9c6ee456f465bb913b0fed4e6390161f0"} err="failed to get container status \"86acdab2983506e5442611ad9ce6fdc9c6ee456f465bb913b0fed4e6390161f0\": rpc error: code = Unknown desc = Error response from daemon: No such container: 86acdab2983506e5442611ad9ce6fdc9c6ee456f465bb913b0fed4e6390161f0"
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.822136 2451 scope.go:117] "RemoveContainer" containerID="6d1e4de06db8f2f6fabcfdf8fb7586505ba8cc6f134c275294b36f4e8ad49fac"
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.852224 2451 scope.go:117] "RemoveContainer" containerID="6d1e4de06db8f2f6fabcfdf8fb7586505ba8cc6f134c275294b36f4e8ad49fac"
Sep 15 18:10:31 addons-924081 kubelet[2451]: E0915 18:10:31.853282 2451 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 6d1e4de06db8f2f6fabcfdf8fb7586505ba8cc6f134c275294b36f4e8ad49fac" containerID="6d1e4de06db8f2f6fabcfdf8fb7586505ba8cc6f134c275294b36f4e8ad49fac"
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.853328 2451 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"6d1e4de06db8f2f6fabcfdf8fb7586505ba8cc6f134c275294b36f4e8ad49fac"} err="failed to get container status \"6d1e4de06db8f2f6fabcfdf8fb7586505ba8cc6f134c275294b36f4e8ad49fac\": rpc error: code = Unknown desc = Error response from daemon: No such container: 6d1e4de06db8f2f6fabcfdf8fb7586505ba8cc6f134c275294b36f4e8ad49fac"
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.853358 2451 scope.go:117] "RemoveContainer" containerID="99c979201ef97c6c87baf1c857f890b602203a87ec0726983cbb938e98f2ce6c"
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.936485 2451 scope.go:117] "RemoveContainer" containerID="99c979201ef97c6c87baf1c857f890b602203a87ec0726983cbb938e98f2ce6c"
Sep 15 18:10:31 addons-924081 kubelet[2451]: E0915 18:10:31.937968 2451 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 99c979201ef97c6c87baf1c857f890b602203a87ec0726983cbb938e98f2ce6c" containerID="99c979201ef97c6c87baf1c857f890b602203a87ec0726983cbb938e98f2ce6c"
Sep 15 18:10:31 addons-924081 kubelet[2451]: I0915 18:10:31.938022 2451 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"99c979201ef97c6c87baf1c857f890b602203a87ec0726983cbb938e98f2ce6c"} err="failed to get container status \"99c979201ef97c6c87baf1c857f890b602203a87ec0726983cbb938e98f2ce6c\": rpc error: code = Unknown desc = Error response from daemon: No such container: 99c979201ef97c6c87baf1c857f890b602203a87ec0726983cbb938e98f2ce6c"
==> storage-provisioner [182d5cd1be62] <==
I0915 17:57:56.326860 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0915 17:57:56.441814 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0915 17:57:56.441871 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0915 17:57:56.532148 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0915 17:57:56.532345 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-924081_bb71074d-2180-4a47-a26d-d7386c9afdc8!
I0915 17:57:56.532415 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"0aa87fe3-e58d-4a23-bc68-a3aee9e3d268", APIVersion:"v1", ResourceVersion:"623", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-924081_bb71074d-2180-4a47-a26d-d7386c9afdc8 became leader
I0915 17:57:56.729148 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-924081_bb71074d-2180-4a47-a26d-d7386c9afdc8!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-924081 -n addons-924081
helpers_test.go:261: (dbg) Run: kubectl --context addons-924081 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox hello-world-app-55bf9c44b4-xz47d
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-924081 describe pod busybox hello-world-app-55bf9c44b4-xz47d
helpers_test.go:282: (dbg) kubectl --context addons-924081 describe pod busybox hello-world-app-55bf9c44b4-xz47d:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-924081/192.168.49.2
Start Time: Sun, 15 Sep 2024 18:01:16 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.28
IPs:
IP: 10.244.0.28
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-x6f9c (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-x6f9c:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m17s default-scheduler Successfully assigned default/busybox to addons-924081
Normal Pulling 7m46s (x4 over 9m16s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m46s (x4 over 9m16s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m46s (x4 over 9m16s) kubelet Error: ErrImagePull
Warning Failed 7m34s (x6 over 9m16s) kubelet Error: ImagePullBackOff
Normal BackOff 4m6s (x21 over 9m16s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Name: hello-world-app-55bf9c44b4-xz47d
Namespace: default
Priority: 0
Service Account: default
Node: addons-924081/192.168.49.2
Start Time: Sun, 15 Sep 2024 18:10:30 +0000
Labels: app=hello-world-app
pod-template-hash=55bf9c44b4
Annotations: <none>
Status: Pending
IP:
IPs: <none>
Controlled By: ReplicaSet/hello-world-app-55bf9c44b4
Containers:
hello-world-app:
Container ID:
Image: docker.io/kicbase/echo-server:1.0
Image ID:
Port: 8080/TCP
Host Port: 0/TCP
State: Waiting
Reason: ContainerCreating
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-t77dd (ro)
Conditions:
Type Status
PodReadyToStartContainers False
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-t77dd:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 3s default-scheduler Successfully assigned default/hello-world-app-55bf9c44b4-xz47d to addons-924081
Normal Pulling 2s kubelet Pulling image "docker.io/kicbase/echo-server:1.0"
Normal Pulled 1s kubelet Successfully pulled image "docker.io/kicbase/echo-server:1.0" in 1.898s (1.898s including waiting). Image size: 4939776 bytes.
Normal Created 1s kubelet Created container hello-world-app
Normal Started 0s kubelet Started container hello-world-app
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (74.17s)