=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:332: registry stabilized in 2.128345ms
addons_test.go:334: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-6fb4cdfc84-g5pxq" [19051672-048d-4f1c-8814-35c5fa1de42e] Running
addons_test.go:334: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.00371888s
addons_test.go:337: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-dsb8t" [ad21aca5-affd-4e4b-9d2e-487316ad11de] Running
addons_test.go:337: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.003386371s
addons_test.go:342: (dbg) Run: kubectl --context addons-271785 delete po -l run=registry-test --now
addons_test.go:347: (dbg) Run: kubectl --context addons-271785 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:347: (dbg) Non-zero exit: kubectl --context addons-271785 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.077144424s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:349: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-271785 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:353: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:361: (dbg) Run: out/minikube-linux-amd64 -p addons-271785 ip
2024/09/09 10:57:38 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:390: (dbg) Run: out/minikube-linux-amd64 -p addons-271785 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-271785
helpers_test.go:235: (dbg) docker inspect addons-271785:
-- stdout --
[
{
"Id": "b659c935a47ab9c53599c159c1feca0b7872c0f8abb12423a7213e56febf5056",
"Created": "2024-09-09T10:44:36.919665652Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 17527,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-09T10:44:37.044870401Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:aeed0e1d4642008f872cbedd0f6935323c1e533683c40e800e0b01d063d11a3c",
"ResolvConfPath": "/var/lib/docker/containers/b659c935a47ab9c53599c159c1feca0b7872c0f8abb12423a7213e56febf5056/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/b659c935a47ab9c53599c159c1feca0b7872c0f8abb12423a7213e56febf5056/hostname",
"HostsPath": "/var/lib/docker/containers/b659c935a47ab9c53599c159c1feca0b7872c0f8abb12423a7213e56febf5056/hosts",
"LogPath": "/var/lib/docker/containers/b659c935a47ab9c53599c159c1feca0b7872c0f8abb12423a7213e56febf5056/b659c935a47ab9c53599c159c1feca0b7872c0f8abb12423a7213e56febf5056-json.log",
"Name": "/addons-271785",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-271785:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-271785",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/217e89f4433d7afec33f368664053c79375ecf7507fa6b06996d74c1c2f6f18a-init/diff:/var/lib/docker/overlay2/f89feb9d9bf85ad5dca6b2eeccfb67947d4725a0c38e64ceddf079e267f149b3/diff",
"MergedDir": "/var/lib/docker/overlay2/217e89f4433d7afec33f368664053c79375ecf7507fa6b06996d74c1c2f6f18a/merged",
"UpperDir": "/var/lib/docker/overlay2/217e89f4433d7afec33f368664053c79375ecf7507fa6b06996d74c1c2f6f18a/diff",
"WorkDir": "/var/lib/docker/overlay2/217e89f4433d7afec33f368664053c79375ecf7507fa6b06996d74c1c2f6f18a/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-271785",
"Source": "/var/lib/docker/volumes/addons-271785/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-271785",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-271785",
"name.minikube.sigs.k8s.io": "addons-271785",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "a6027eb71cbd8578dbef8bcb98d2c45e8dbbf60d2f25a52e060bfe44c5641404",
"SandboxKey": "/var/run/docker/netns/a6027eb71cbd",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-271785": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "d8e492975c4903b17a21a80a8562f31fedb32d6ca1ad1a358d25444bc4d69f7e",
"EndpointID": "8b089c9e663cf0cf10f18cb325596a6b178ff14e0e599ffc197847e8aa4357b4",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-271785",
"b659c935a47a"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-271785 -n addons-271785
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-271785 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-docker-180363 | download-docker-180363 | jenkins | v1.34.0 | 09 Sep 24 10:44 UTC | 09 Sep 24 10:44 UTC |
| start | --download-only -p | binary-mirror-876583 | jenkins | v1.34.0 | 09 Sep 24 10:44 UTC | |
| | binary-mirror-876583 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:43487 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-876583 | binary-mirror-876583 | jenkins | v1.34.0 | 09 Sep 24 10:44 UTC | 09 Sep 24 10:44 UTC |
| addons | enable dashboard -p | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:44 UTC | |
| | addons-271785 | | | | | |
| addons | disable dashboard -p | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:44 UTC | |
| | addons-271785 | | | | | |
| start | -p addons-271785 --wait=true | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:44 UTC | 09 Sep 24 10:47 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | addons-271785 addons disable | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:48 UTC | 09 Sep 24 10:48 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | addons-271785 addons | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:56 UTC | 09 Sep 24 10:56 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-271785 addons disable | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:56 UTC | 09 Sep 24 10:56 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | addons-271785 addons disable | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:56 UTC | 09 Sep 24 10:56 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:56 UTC | 09 Sep 24 10:56 UTC |
| | -p addons-271785 | | | | | |
| addons | disable cloud-spanner -p | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:56 UTC | 09 Sep 24 10:56 UTC |
| | addons-271785 | | | | | |
| addons | enable headlamp | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:56 UTC | 09 Sep 24 10:56 UTC |
| | -p addons-271785 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-271785 ssh cat | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:56 UTC | 09 Sep 24 10:56 UTC |
| | /opt/local-path-provisioner/pvc-88f0cab2-ac8e-4b40-842d-f0e3d852d155_default_test-pvc/file1 | | | | | |
| addons | addons-271785 addons disable | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:56 UTC | 09 Sep 24 10:57 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-271785 addons disable | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:57 UTC | 09 Sep 24 10:57 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-271785 addons | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:57 UTC | 09 Sep 24 10:57 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:57 UTC | 09 Sep 24 10:57 UTC |
| | addons-271785 | | | | | |
| addons | addons-271785 addons | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:57 UTC | 09 Sep 24 10:57 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-271785 ssh curl -s | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:57 UTC | 09 Sep 24 10:57 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-271785 ip | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:57 UTC | 09 Sep 24 10:57 UTC |
| addons | addons-271785 addons disable | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:57 UTC | 09 Sep 24 10:57 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-271785 addons disable | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:57 UTC | 09 Sep 24 10:57 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| ip | addons-271785 ip | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:57 UTC | 09 Sep 24 10:57 UTC |
| addons | addons-271785 addons disable | addons-271785 | jenkins | v1.34.0 | 09 Sep 24 10:57 UTC | 09 Sep 24 10:57 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/09 10:44:15
Running on machine: ubuntu-20-agent
Binary: Built with gc go1.22.5 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0909 10:44:15.522778 16785 out.go:345] Setting OutFile to fd 1 ...
I0909 10:44:15.522928 16785 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0909 10:44:15.522939 16785 out.go:358] Setting ErrFile to fd 2...
I0909 10:44:15.522945 16785 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0909 10:44:15.523168 16785 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19584-8635/.minikube/bin
I0909 10:44:15.523819 16785 out.go:352] Setting JSON to false
I0909 10:44:15.524700 16785 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent","uptime":1588,"bootTime":1725877067,"procs":169,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1067-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0909 10:44:15.524761 16785 start.go:139] virtualization: kvm guest
I0909 10:44:15.528000 16785 out.go:177] * [addons-271785] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0909 10:44:15.529352 16785 notify.go:220] Checking for updates...
I0909 10:44:15.529377 16785 out.go:177] - MINIKUBE_LOCATION=19584
I0909 10:44:15.530712 16785 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0909 10:44:15.532070 16785 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19584-8635/kubeconfig
I0909 10:44:15.533468 16785 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19584-8635/.minikube
I0909 10:44:15.534965 16785 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0909 10:44:15.536049 16785 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0909 10:44:15.537224 16785 driver.go:394] Setting default libvirt URI to qemu:///system
I0909 10:44:15.557700 16785 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
I0909 10:44:15.557802 16785 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0909 10:44:15.600757 16785 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-09 10:44:15.591919187 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1067-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:472731909fa34bd7bc9c087e4c27943f9835f111 Expected:472731909fa34bd7bc9c087e4c27943f9835f111} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerError
s:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0909 10:44:15.600857 16785 docker.go:307] overlay module found
I0909 10:44:15.602465 16785 out.go:177] * Using the docker driver based on user configuration
I0909 10:44:15.603516 16785 start.go:297] selected driver: docker
I0909 10:44:15.603526 16785 start.go:901] validating driver "docker" against <nil>
I0909 10:44:15.603535 16785 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0909 10:44:15.604258 16785 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0909 10:44:15.651961 16785 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-09 10:44:15.643582469 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1067-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:472731909fa34bd7bc9c087e4c27943f9835f111 Expected:472731909fa34bd7bc9c087e4c27943f9835f111} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerError
s:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0909 10:44:15.652122 16785 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0909 10:44:15.652331 16785 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0909 10:44:15.653953 16785 out.go:177] * Using Docker driver with root privileges
I0909 10:44:15.655232 16785 cni.go:84] Creating CNI manager for ""
I0909 10:44:15.655255 16785 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0909 10:44:15.655268 16785 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0909 10:44:15.655330 16785 start.go:340] cluster config:
{Name:addons-271785 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-271785 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: Netwo
rkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: Aut
oPauseInterval:1m0s}
I0909 10:44:15.656699 16785 out.go:177] * Starting "addons-271785" primary control-plane node in "addons-271785" cluster
I0909 10:44:15.657859 16785 cache.go:121] Beginning downloading kic base image for docker with docker
I0909 10:44:15.658998 16785 out.go:177] * Pulling base image v0.0.45 ...
I0909 10:44:15.660066 16785 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0909 10:44:15.660092 16785 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19584-8635/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4
I0909 10:44:15.660103 16785 cache.go:56] Caching tarball of preloaded images
I0909 10:44:15.660158 16785 image.go:79] Checking for gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 in local docker daemon
I0909 10:44:15.660221 16785 preload.go:172] Found /home/jenkins/minikube-integration/19584-8635/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0909 10:44:15.660235 16785 cache.go:59] Finished verifying existence of preloaded tar for v1.31.0 on docker
I0909 10:44:15.660563 16785 profile.go:143] Saving config to /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/config.json ...
I0909 10:44:15.660622 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/config.json: {Name:mkb74be3b59a4fdd9e36d3b7a352daaaec2eb359 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:15.676210 16785 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 to local cache
I0909 10:44:15.676302 16785 image.go:63] Checking for gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 in local cache directory
I0909 10:44:15.676321 16785 image.go:66] Found gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 in local cache directory, skipping pull
I0909 10:44:15.676330 16785 image.go:135] gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 exists in cache, skipping pull
I0909 10:44:15.676337 16785 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 as a tarball
I0909 10:44:15.676344 16785 cache.go:162] Loading gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 from local cache
I0909 10:44:27.607353 16785 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 from cached tarball
I0909 10:44:27.607391 16785 cache.go:194] Successfully downloaded all kic artifacts
I0909 10:44:27.607429 16785 start.go:360] acquireMachinesLock for addons-271785: {Name:mk4a4f218a8b2d1a95515c42d18e18c0b87b1ced Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0909 10:44:27.607521 16785 start.go:364] duration metric: took 73.784µs to acquireMachinesLock for "addons-271785"
I0909 10:44:27.607542 16785 start.go:93] Provisioning new machine with config: &{Name:addons-271785 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-271785 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[
] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: Sock
etVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0909 10:44:27.607631 16785 start.go:125] createHost starting for "" (driver="docker")
I0909 10:44:27.609149 16785 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0909 10:44:27.609346 16785 start.go:159] libmachine.API.Create for "addons-271785" (driver="docker")
I0909 10:44:27.609379 16785 client.go:168] LocalClient.Create starting
I0909 10:44:27.609481 16785 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19584-8635/.minikube/certs/ca.pem
I0909 10:44:27.804430 16785 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19584-8635/.minikube/certs/cert.pem
I0909 10:44:27.953718 16785 cli_runner.go:164] Run: docker network inspect addons-271785 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0909 10:44:27.969011 16785 cli_runner.go:211] docker network inspect addons-271785 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0909 10:44:27.969087 16785 network_create.go:284] running [docker network inspect addons-271785] to gather additional debugging logs...
I0909 10:44:27.969103 16785 cli_runner.go:164] Run: docker network inspect addons-271785
W0909 10:44:27.983687 16785 cli_runner.go:211] docker network inspect addons-271785 returned with exit code 1
I0909 10:44:27.983713 16785 network_create.go:287] error running [docker network inspect addons-271785]: docker network inspect addons-271785: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-271785 not found
I0909 10:44:27.983723 16785 network_create.go:289] output of [docker network inspect addons-271785]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-271785 not found
** /stderr **
I0909 10:44:27.983810 16785 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0909 10:44:27.998813 16785 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001b0a910}
I0909 10:44:27.998851 16785 network_create.go:124] attempt to create docker network addons-271785 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0909 10:44:27.998887 16785 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-271785 addons-271785
I0909 10:44:28.055451 16785 network_create.go:108] docker network addons-271785 192.168.49.0/24 created
I0909 10:44:28.055481 16785 kic.go:121] calculated static IP "192.168.49.2" for the "addons-271785" container
I0909 10:44:28.055535 16785 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0909 10:44:28.069959 16785 cli_runner.go:164] Run: docker volume create addons-271785 --label name.minikube.sigs.k8s.io=addons-271785 --label created_by.minikube.sigs.k8s.io=true
I0909 10:44:28.085741 16785 oci.go:103] Successfully created a docker volume addons-271785
I0909 10:44:28.085804 16785 cli_runner.go:164] Run: docker run --rm --name addons-271785-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-271785 --entrypoint /usr/bin/test -v addons-271785:/var gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 -d /var/lib
I0909 10:44:32.989476 16785 cli_runner.go:217] Completed: docker run --rm --name addons-271785-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-271785 --entrypoint /usr/bin/test -v addons-271785:/var gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 -d /var/lib: (4.903633286s)
I0909 10:44:32.989506 16785 oci.go:107] Successfully prepared a docker volume addons-271785
I0909 10:44:32.989522 16785 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0909 10:44:32.989539 16785 kic.go:194] Starting extracting preloaded images to volume ...
I0909 10:44:32.989583 16785 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19584-8635/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-271785:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 -I lz4 -xf /preloaded.tar -C /extractDir
I0909 10:44:36.856019 16785 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19584-8635/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-271785:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 -I lz4 -xf /preloaded.tar -C /extractDir: (3.866398431s)
I0909 10:44:36.856050 16785 kic.go:203] duration metric: took 3.866508055s to extract preloaded images to volume ...
W0909 10:44:36.856177 16785 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0909 10:44:36.856287 16785 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0909 10:44:36.905606 16785 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-271785 --name addons-271785 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-271785 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-271785 --network addons-271785 --ip 192.168.49.2 --volume addons-271785:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85
I0909 10:44:37.214346 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Running}}
I0909 10:44:37.233687 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:37.251040 16785 cli_runner.go:164] Run: docker exec addons-271785 stat /var/lib/dpkg/alternatives/iptables
I0909 10:44:37.291640 16785 oci.go:144] the created container "addons-271785" has a running status.
I0909 10:44:37.291674 16785 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa...
I0909 10:44:37.464178 16785 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0909 10:44:37.489507 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:37.508271 16785 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0909 10:44:37.508296 16785 kic_runner.go:114] Args: [docker exec --privileged addons-271785 chown docker:docker /home/docker/.ssh/authorized_keys]
I0909 10:44:37.564428 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:37.585511 16785 machine.go:93] provisionDockerMachine start ...
I0909 10:44:37.585611 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:37.604185 16785 main.go:141] libmachine: Using SSH client type: native
I0909 10:44:37.604399 16785 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0909 10:44:37.604417 16785 main.go:141] libmachine: About to run SSH command:
hostname
I0909 10:44:37.795598 16785 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-271785
I0909 10:44:37.795626 16785 ubuntu.go:169] provisioning hostname "addons-271785"
I0909 10:44:37.795677 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:37.814070 16785 main.go:141] libmachine: Using SSH client type: native
I0909 10:44:37.814238 16785 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0909 10:44:37.814252 16785 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-271785 && echo "addons-271785" | sudo tee /etc/hostname
I0909 10:44:37.950451 16785 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-271785
I0909 10:44:37.950517 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:37.967658 16785 main.go:141] libmachine: Using SSH client type: native
I0909 10:44:37.967822 16785 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0909 10:44:37.967839 16785 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-271785' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-271785/g' /etc/hosts;
else
echo '127.0.1.1 addons-271785' | sudo tee -a /etc/hosts;
fi
fi
I0909 10:44:38.092180 16785 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0909 10:44:38.092208 16785 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19584-8635/.minikube CaCertPath:/home/jenkins/minikube-integration/19584-8635/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19584-8635/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19584-8635/.minikube}
I0909 10:44:38.092235 16785 ubuntu.go:177] setting up certificates
I0909 10:44:38.092246 16785 provision.go:84] configureAuth start
I0909 10:44:38.092288 16785 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-271785
I0909 10:44:38.108128 16785 provision.go:143] copyHostCerts
I0909 10:44:38.108207 16785 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19584-8635/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19584-8635/.minikube/ca.pem (1078 bytes)
I0909 10:44:38.108326 16785 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19584-8635/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19584-8635/.minikube/cert.pem (1123 bytes)
I0909 10:44:38.108406 16785 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19584-8635/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19584-8635/.minikube/key.pem (1679 bytes)
I0909 10:44:38.108480 16785 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19584-8635/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19584-8635/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19584-8635/.minikube/certs/ca-key.pem org=jenkins.addons-271785 san=[127.0.0.1 192.168.49.2 addons-271785 localhost minikube]
I0909 10:44:38.211639 16785 provision.go:177] copyRemoteCerts
I0909 10:44:38.211704 16785 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0909 10:44:38.211751 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:38.227730 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:38.316590 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0909 10:44:38.336464 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0909 10:44:38.356267 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0909 10:44:38.375936 16785 provision.go:87] duration metric: took 283.678399ms to configureAuth
I0909 10:44:38.375964 16785 ubuntu.go:193] setting minikube options for container-runtime
I0909 10:44:38.376135 16785 config.go:182] Loaded profile config "addons-271785": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0909 10:44:38.376188 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:38.392440 16785 main.go:141] libmachine: Using SSH client type: native
I0909 10:44:38.392638 16785 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0909 10:44:38.392654 16785 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0909 10:44:38.512665 16785 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0909 10:44:38.512687 16785 ubuntu.go:71] root file system type: overlay
I0909 10:44:38.512807 16785 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0909 10:44:38.512856 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:38.529320 16785 main.go:141] libmachine: Using SSH client type: native
I0909 10:44:38.529492 16785 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0909 10:44:38.529548 16785 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0909 10:44:38.658095 16785 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0909 10:44:38.658189 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:38.674112 16785 main.go:141] libmachine: Using SSH client type: native
I0909 10:44:38.674276 16785 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0909 10:44:38.674292 16785 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0909 10:44:39.320407 16785 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-08-27 14:13:43.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-09 10:44:38.656005256 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0909 10:44:39.320441 16785 machine.go:96] duration metric: took 1.734902837s to provisionDockerMachine
I0909 10:44:39.320452 16785 client.go:171] duration metric: took 11.711065768s to LocalClient.Create
I0909 10:44:39.320466 16785 start.go:167] duration metric: took 11.711120564s to libmachine.API.Create "addons-271785"
I0909 10:44:39.320473 16785 start.go:293] postStartSetup for "addons-271785" (driver="docker")
I0909 10:44:39.320483 16785 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0909 10:44:39.320532 16785 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0909 10:44:39.320588 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:39.336514 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:39.424662 16785 ssh_runner.go:195] Run: cat /etc/os-release
I0909 10:44:39.427344 16785 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0909 10:44:39.427373 16785 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0909 10:44:39.427392 16785 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0909 10:44:39.427406 16785 info.go:137] Remote host: Ubuntu 22.04.4 LTS
I0909 10:44:39.427422 16785 filesync.go:126] Scanning /home/jenkins/minikube-integration/19584-8635/.minikube/addons for local assets ...
I0909 10:44:39.427475 16785 filesync.go:126] Scanning /home/jenkins/minikube-integration/19584-8635/.minikube/files for local assets ...
I0909 10:44:39.427505 16785 start.go:296] duration metric: took 107.02607ms for postStartSetup
I0909 10:44:39.427818 16785 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-271785
I0909 10:44:39.443689 16785 profile.go:143] Saving config to /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/config.json ...
I0909 10:44:39.443913 16785 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0909 10:44:39.443950 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:39.459207 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:39.545012 16785 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0909 10:44:39.548798 16785 start.go:128] duration metric: took 11.941155404s to createHost
I0909 10:44:39.548821 16785 start.go:83] releasing machines lock for "addons-271785", held for 11.941289683s
I0909 10:44:39.548880 16785 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-271785
I0909 10:44:39.564556 16785 ssh_runner.go:195] Run: cat /version.json
I0909 10:44:39.564628 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:39.564637 16785 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0909 10:44:39.564702 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:39.580503 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:39.581907 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:39.737240 16785 ssh_runner.go:195] Run: systemctl --version
I0909 10:44:39.741033 16785 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0909 10:44:39.744673 16785 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0909 10:44:39.764905 16785 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0909 10:44:39.764961 16785 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0909 10:44:39.787509 16785 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0909 10:44:39.787530 16785 start.go:495] detecting cgroup driver to use...
I0909 10:44:39.787561 16785 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0909 10:44:39.787677 16785 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0909 10:44:39.800966 16785 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0909 10:44:39.809035 16785 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0909 10:44:39.817030 16785 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0909 10:44:39.817101 16785 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0909 10:44:39.825116 16785 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0909 10:44:39.832936 16785 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0909 10:44:39.840510 16785 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0909 10:44:39.848273 16785 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0909 10:44:39.856093 16785 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0909 10:44:39.863763 16785 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0909 10:44:39.871650 16785 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0909 10:44:39.879305 16785 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0909 10:44:39.885942 16785 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0909 10:44:39.892624 16785 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0909 10:44:39.970045 16785 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0909 10:44:40.057535 16785 start.go:495] detecting cgroup driver to use...
I0909 10:44:40.057583 16785 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0909 10:44:40.057629 16785 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0909 10:44:40.068071 16785 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0909 10:44:40.068139 16785 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0909 10:44:40.078368 16785 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0909 10:44:40.093454 16785 ssh_runner.go:195] Run: which cri-dockerd
I0909 10:44:40.096492 16785 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0909 10:44:40.104595 16785 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0909 10:44:40.119752 16785 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0909 10:44:40.196549 16785 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0909 10:44:40.288130 16785 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0909 10:44:40.288238 16785 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0909 10:44:40.303509 16785 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0909 10:44:40.381331 16785 ssh_runner.go:195] Run: sudo systemctl restart docker
I0909 10:44:40.618626 16785 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0909 10:44:40.628678 16785 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0909 10:44:40.638151 16785 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0909 10:44:40.709246 16785 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0909 10:44:40.782286 16785 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0909 10:44:40.854128 16785 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0909 10:44:40.865333 16785 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0909 10:44:40.874355 16785 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0909 10:44:40.954372 16785 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0909 10:44:41.009242 16785 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0909 10:44:41.009313 16785 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0909 10:44:41.012671 16785 start.go:563] Will wait 60s for crictl version
I0909 10:44:41.012722 16785 ssh_runner.go:195] Run: which crictl
I0909 10:44:41.015492 16785 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0909 10:44:41.044526 16785 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.2.0
RuntimeApiVersion: v1
I0909 10:44:41.044598 16785 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0909 10:44:41.066358 16785 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0909 10:44:41.090602 16785 out.go:235] * Preparing Kubernetes v1.31.0 on Docker 27.2.0 ...
I0909 10:44:41.090671 16785 cli_runner.go:164] Run: docker network inspect addons-271785 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0909 10:44:41.105742 16785 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0909 10:44:41.108999 16785 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0909 10:44:41.118405 16785 kubeadm.go:883] updating cluster {Name:addons-271785 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-271785 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] D
NSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnet
ClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0909 10:44:41.118501 16785 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0909 10:44:41.118537 16785 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0909 10:44:41.135979 16785 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-controller-manager:v1.31.0
registry.k8s.io/kube-scheduler:v1.31.0
registry.k8s.io/kube-apiserver:v1.31.0
registry.k8s.io/kube-proxy:v1.31.0
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
registry.k8s.io/coredns/coredns:v1.11.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0909 10:44:41.135999 16785 docker.go:615] Images already preloaded, skipping extraction
I0909 10:44:41.136049 16785 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0909 10:44:41.152249 16785 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-controller-manager:v1.31.0
registry.k8s.io/kube-scheduler:v1.31.0
registry.k8s.io/kube-apiserver:v1.31.0
registry.k8s.io/kube-proxy:v1.31.0
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
registry.k8s.io/coredns/coredns:v1.11.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0909 10:44:41.152269 16785 cache_images.go:84] Images are preloaded, skipping loading
I0909 10:44:41.152289 16785 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.0 docker true true} ...
I0909 10:44:41.152393 16785 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.0/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-271785 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.0 ClusterName:addons-271785 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0909 10:44:41.152438 16785 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0909 10:44:41.193174 16785 cni.go:84] Creating CNI manager for ""
I0909 10:44:41.193196 16785 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0909 10:44:41.193217 16785 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0909 10:44:41.193235 16785 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.0 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-271785 NodeName:addons-271785 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0909 10:44:41.193352 16785 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-271785"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.0
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0909 10:44:41.193427 16785 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.0
I0909 10:44:41.200967 16785 binaries.go:44] Found k8s binaries, skipping transfer
I0909 10:44:41.201021 16785 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0909 10:44:41.208160 16785 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0909 10:44:41.222628 16785 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0909 10:44:41.236674 16785 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0909 10:44:41.251378 16785 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0909 10:44:41.254225 16785 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0909 10:44:41.263243 16785 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0909 10:44:41.335521 16785 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0909 10:44:41.347198 16785 certs.go:68] Setting up /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785 for IP: 192.168.49.2
I0909 10:44:41.347217 16785 certs.go:194] generating shared ca certs ...
I0909 10:44:41.347231 16785 certs.go:226] acquiring lock for ca certs: {Name:mk2360cf7fa1bb5fb294939d08b9d4b496d4efcc Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:41.347348 16785 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19584-8635/.minikube/ca.key
I0909 10:44:41.512972 16785 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19584-8635/.minikube/ca.crt ...
I0909 10:44:41.512998 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/ca.crt: {Name:mk88d0f9b2cb901083f6a448fea938fdbaa0d8cf Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:41.513150 16785 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19584-8635/.minikube/ca.key ...
I0909 10:44:41.513160 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/ca.key: {Name:mk875f3e16cd09eaea01032711ec3f531dc6b622 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:41.513227 16785 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19584-8635/.minikube/proxy-client-ca.key
I0909 10:44:41.811128 16785 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19584-8635/.minikube/proxy-client-ca.crt ...
I0909 10:44:41.811160 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/proxy-client-ca.crt: {Name:mk737528587ea1f5caaab01eb78f2719c027692b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:41.811340 16785 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19584-8635/.minikube/proxy-client-ca.key ...
I0909 10:44:41.811356 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/proxy-client-ca.key: {Name:mkd183719606887aad007d2a4b2c2ccae5de918f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:41.811459 16785 certs.go:256] generating profile certs ...
I0909 10:44:41.811530 16785 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/client.key
I0909 10:44:41.811548 16785 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/client.crt with IP's: []
I0909 10:44:41.926408 16785 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/client.crt ...
I0909 10:44:41.926437 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/client.crt: {Name:mk1c994c64ac3cf5e7745b66bb18fee9265eca18 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:41.926602 16785 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/client.key ...
I0909 10:44:41.926617 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/client.key: {Name:mk9196b6a7a2572d0448feb6e65f57c8f27d6c98 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:41.926734 16785 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.key.3e96fd18
I0909 10:44:41.926758 16785 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.crt.3e96fd18 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0909 10:44:42.120731 16785 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.crt.3e96fd18 ...
I0909 10:44:42.120762 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.crt.3e96fd18: {Name:mk9ab0260d612b029bcb2ac113e83dab71c3cdd6 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:42.120933 16785 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.key.3e96fd18 ...
I0909 10:44:42.120952 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.key.3e96fd18: {Name:mk27eb5d25c718ddbd0ab42f77eef30523fca64b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:42.121049 16785 certs.go:381] copying /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.crt.3e96fd18 -> /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.crt
I0909 10:44:42.121140 16785 certs.go:385] copying /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.key.3e96fd18 -> /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.key
I0909 10:44:42.121212 16785 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/proxy-client.key
I0909 10:44:42.121240 16785 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/proxy-client.crt with IP's: []
I0909 10:44:42.455357 16785 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/proxy-client.crt ...
I0909 10:44:42.455391 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/proxy-client.crt: {Name:mk0a021f98b1c8f3e4e3f39e1d772537089a22cf Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:42.455581 16785 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/proxy-client.key ...
I0909 10:44:42.455597 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/proxy-client.key: {Name:mka71d079ae74c19ce1899466631684ab3eafe05 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:42.455796 16785 certs.go:484] found cert: /home/jenkins/minikube-integration/19584-8635/.minikube/certs/ca-key.pem (1675 bytes)
I0909 10:44:42.455840 16785 certs.go:484] found cert: /home/jenkins/minikube-integration/19584-8635/.minikube/certs/ca.pem (1078 bytes)
I0909 10:44:42.455873 16785 certs.go:484] found cert: /home/jenkins/minikube-integration/19584-8635/.minikube/certs/cert.pem (1123 bytes)
I0909 10:44:42.455909 16785 certs.go:484] found cert: /home/jenkins/minikube-integration/19584-8635/.minikube/certs/key.pem (1679 bytes)
I0909 10:44:42.456526 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0909 10:44:42.477301 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0909 10:44:42.496876 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0909 10:44:42.516814 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0909 10:44:42.536766 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0909 10:44:42.556378 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0909 10:44:42.575787 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0909 10:44:42.595252 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/profiles/addons-271785/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0909 10:44:42.614331 16785 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19584-8635/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0909 10:44:42.633911 16785 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0909 10:44:42.648089 16785 ssh_runner.go:195] Run: openssl version
I0909 10:44:42.652658 16785 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0909 10:44:42.660254 16785 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0909 10:44:42.663120 16785 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 9 10:44 /usr/share/ca-certificates/minikubeCA.pem
I0909 10:44:42.663159 16785 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0909 10:44:42.668989 16785 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0909 10:44:42.676459 16785 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0909 10:44:42.679060 16785 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0909 10:44:42.679107 16785 kubeadm.go:392] StartCluster: {Name:addons-271785 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-271785 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSD
omain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetCli
entPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0909 10:44:42.679194 16785 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0909 10:44:42.694624 16785 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0909 10:44:42.701981 16785 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0909 10:44:42.709294 16785 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0909 10:44:42.709340 16785 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0909 10:44:42.716623 16785 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0909 10:44:42.716653 16785 kubeadm.go:157] found existing configuration files:
I0909 10:44:42.716689 16785 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0909 10:44:42.723771 16785 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0909 10:44:42.723827 16785 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0909 10:44:42.730916 16785 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0909 10:44:42.738197 16785 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0909 10:44:42.738254 16785 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0909 10:44:42.745097 16785 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0909 10:44:42.752100 16785 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0909 10:44:42.752142 16785 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0909 10:44:42.758963 16785 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0909 10:44:42.765731 16785 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0909 10:44:42.765780 16785 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0909 10:44:42.772442 16785 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.0:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0909 10:44:42.805808 16785 kubeadm.go:310] [init] Using Kubernetes version: v1.31.0
I0909 10:44:42.805867 16785 kubeadm.go:310] [preflight] Running pre-flight checks
I0909 10:44:42.827012 16785 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0909 10:44:42.827119 16785 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1067-gcp[0m
I0909 10:44:42.827177 16785 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0909 10:44:42.827258 16785 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0909 10:44:42.827307 16785 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0909 10:44:42.827366 16785 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0909 10:44:42.827435 16785 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0909 10:44:42.827503 16785 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0909 10:44:42.827573 16785 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0909 10:44:42.827639 16785 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0909 10:44:42.827735 16785 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0909 10:44:42.827805 16785 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0909 10:44:42.876643 16785 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0909 10:44:42.876820 16785 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0909 10:44:42.876951 16785 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0909 10:44:42.886858 16785 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0909 10:44:42.889705 16785 out.go:235] - Generating certificates and keys ...
I0909 10:44:42.889819 16785 kubeadm.go:310] [certs] Using existing ca certificate authority
I0909 10:44:42.889902 16785 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0909 10:44:43.020429 16785 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0909 10:44:43.072088 16785 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0909 10:44:43.175075 16785 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0909 10:44:43.259521 16785 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0909 10:44:43.307625 16785 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0909 10:44:43.307761 16785 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-271785 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0909 10:44:43.397470 16785 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0909 10:44:43.397650 16785 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-271785 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0909 10:44:43.483485 16785 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0909 10:44:43.643546 16785 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0909 10:44:43.729533 16785 kubeadm.go:310] [certs] Generating "sa" key and public key
I0909 10:44:43.729601 16785 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0909 10:44:44.073002 16785 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0909 10:44:44.209564 16785 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0909 10:44:44.297501 16785 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0909 10:44:44.805737 16785 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0909 10:44:44.903940 16785 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0909 10:44:44.904554 16785 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0909 10:44:44.906921 16785 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0909 10:44:44.909032 16785 out.go:235] - Booting up control plane ...
I0909 10:44:44.909167 16785 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0909 10:44:44.909277 16785 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0909 10:44:44.909376 16785 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0909 10:44:44.917482 16785 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0909 10:44:44.922146 16785 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0909 10:44:44.922202 16785 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0909 10:44:45.004711 16785 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0909 10:44:45.004812 16785 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0909 10:44:46.005916 16785 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.00128921s
I0909 10:44:46.006064 16785 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0909 10:44:50.007585 16785 kubeadm.go:310] [api-check] The API server is healthy after 4.001671855s
I0909 10:44:50.018887 16785 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0909 10:44:50.028984 16785 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0909 10:44:50.044054 16785 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0909 10:44:50.044222 16785 kubeadm.go:310] [mark-control-plane] Marking the node addons-271785 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0909 10:44:50.050781 16785 kubeadm.go:310] [bootstrap-token] Using token: 1fluin.246gyo3umj6g529d
I0909 10:44:50.051977 16785 out.go:235] - Configuring RBAC rules ...
I0909 10:44:50.052097 16785 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0909 10:44:50.054765 16785 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0909 10:44:50.059534 16785 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0909 10:44:50.061634 16785 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0909 10:44:50.063678 16785 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0909 10:44:50.066465 16785 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0909 10:44:50.413907 16785 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0909 10:44:50.831506 16785 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0909 10:44:51.413488 16785 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0909 10:44:51.414299 16785 kubeadm.go:310]
I0909 10:44:51.414380 16785 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0909 10:44:51.414391 16785 kubeadm.go:310]
I0909 10:44:51.414478 16785 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0909 10:44:51.414486 16785 kubeadm.go:310]
I0909 10:44:51.414520 16785 kubeadm.go:310] mkdir -p $HOME/.kube
I0909 10:44:51.414592 16785 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0909 10:44:51.414660 16785 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0909 10:44:51.414670 16785 kubeadm.go:310]
I0909 10:44:51.414768 16785 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0909 10:44:51.414783 16785 kubeadm.go:310]
I0909 10:44:51.414852 16785 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0909 10:44:51.414862 16785 kubeadm.go:310]
I0909 10:44:51.414930 16785 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0909 10:44:51.415034 16785 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0909 10:44:51.415130 16785 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0909 10:44:51.415138 16785 kubeadm.go:310]
I0909 10:44:51.415269 16785 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0909 10:44:51.415400 16785 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0909 10:44:51.415415 16785 kubeadm.go:310]
I0909 10:44:51.415553 16785 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 1fluin.246gyo3umj6g529d \
I0909 10:44:51.415703 16785 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:40c376e8c23af1435b6d54a5ad8ad6332034ab7c8b00f2ebf940ca94a37535b9 \
I0909 10:44:51.415733 16785 kubeadm.go:310] --control-plane
I0909 10:44:51.415747 16785 kubeadm.go:310]
I0909 10:44:51.415867 16785 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0909 10:44:51.415876 16785 kubeadm.go:310]
I0909 10:44:51.415990 16785 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token 1fluin.246gyo3umj6g529d \
I0909 10:44:51.416141 16785 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:40c376e8c23af1435b6d54a5ad8ad6332034ab7c8b00f2ebf940ca94a37535b9
I0909 10:44:51.417590 16785 kubeadm.go:310] W0909 10:44:42.803306 1922 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0909 10:44:51.417856 16785 kubeadm.go:310] W0909 10:44:42.803945 1922 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0909 10:44:51.418041 16785 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1067-gcp\n", err: exit status 1
I0909 10:44:51.418145 16785 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0909 10:44:51.418172 16785 cni.go:84] Creating CNI manager for ""
I0909 10:44:51.418187 16785 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0909 10:44:51.419721 16785 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0909 10:44:51.421060 16785 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0909 10:44:51.429203 16785 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0909 10:44:51.443946 16785 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0909 10:44:51.444003 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:51.444021 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-271785 minikube.k8s.io/updated_at=2024_09_09T10_44_51_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=cf17d6b4040a54caaa170f92a048a513bb2a2b0d minikube.k8s.io/name=addons-271785 minikube.k8s.io/primary=true
I0909 10:44:51.522187 16785 ops.go:34] apiserver oom_adj: -16
I0909 10:44:51.522307 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:52.022942 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:52.522527 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:53.022685 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:53.522784 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:54.022509 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:54.522562 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:55.023055 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:55.522537 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:56.022392 16785 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0909 10:44:56.096687 16785 kubeadm.go:1113] duration metric: took 4.652731851s to wait for elevateKubeSystemPrivileges
I0909 10:44:56.096723 16785 kubeadm.go:394] duration metric: took 13.417618087s to StartCluster
I0909 10:44:56.096740 16785 settings.go:142] acquiring lock: {Name:mk36f011397e0c600653f6927921ea8dbea2b461 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:56.096843 16785 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19584-8635/kubeconfig
I0909 10:44:56.097243 16785 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19584-8635/kubeconfig: {Name:mk440c5bd831c615cd310d0b32ed59bfbea69096 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0909 10:44:56.097402 16785 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0909 10:44:56.097425 16785 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0909 10:44:56.097498 16785 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0909 10:44:56.097570 16785 config.go:182] Loaded profile config "addons-271785": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0909 10:44:56.097597 16785 addons.go:69] Setting yakd=true in profile "addons-271785"
I0909 10:44:56.097608 16785 addons.go:69] Setting default-storageclass=true in profile "addons-271785"
I0909 10:44:56.097624 16785 addons.go:69] Setting metrics-server=true in profile "addons-271785"
I0909 10:44:56.097630 16785 addons.go:234] Setting addon yakd=true in "addons-271785"
I0909 10:44:56.097608 16785 addons.go:69] Setting cloud-spanner=true in profile "addons-271785"
I0909 10:44:56.097631 16785 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-271785"
I0909 10:44:56.097656 16785 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-271785"
I0909 10:44:56.097665 16785 addons.go:69] Setting ingress-dns=true in profile "addons-271785"
I0909 10:44:56.097671 16785 addons.go:69] Setting inspektor-gadget=true in profile "addons-271785"
I0909 10:44:56.097684 16785 addons.go:234] Setting addon ingress-dns=true in "addons-271785"
I0909 10:44:56.097689 16785 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-271785"
I0909 10:44:56.097696 16785 addons.go:234] Setting addon inspektor-gadget=true in "addons-271785"
I0909 10:44:56.097700 16785 addons.go:69] Setting volcano=true in profile "addons-271785"
I0909 10:44:56.097712 16785 addons.go:69] Setting helm-tiller=true in profile "addons-271785"
I0909 10:44:56.097717 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.097717 16785 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-271785"
I0909 10:44:56.097725 16785 addons.go:69] Setting volumesnapshots=true in profile "addons-271785"
I0909 10:44:56.097672 16785 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-271785"
I0909 10:44:56.097736 16785 addons.go:69] Setting storage-provisioner=true in profile "addons-271785"
I0909 10:44:56.097742 16785 addons.go:234] Setting addon volumesnapshots=true in "addons-271785"
I0909 10:44:56.097747 16785 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-271785"
I0909 10:44:56.097749 16785 addons.go:234] Setting addon helm-tiller=true in "addons-271785"
I0909 10:44:56.097759 16785 addons.go:234] Setting addon storage-provisioner=true in "addons-271785"
I0909 10:44:56.097765 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.097777 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.097690 16785 addons.go:69] Setting registry=true in profile "addons-271785"
I0909 10:44:56.097787 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.097800 16785 addons.go:234] Setting addon registry=true in "addons-271785"
I0909 10:44:56.097819 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.098037 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.098045 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.098193 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.098207 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.098217 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.098236 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.098247 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.097779 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.097662 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.097647 16785 addons.go:234] Setting addon metrics-server=true in "addons-271785"
I0909 10:44:56.099118 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.097675 16785 addons.go:234] Setting addon cloud-spanner=true in "addons-271785"
I0909 10:44:56.099351 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.097717 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.099604 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.099641 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.099921 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.099946 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.100466 16785 out.go:177] * Verifying Kubernetes components...
I0909 10:44:56.097723 16785 addons.go:234] Setting addon volcano=true in "addons-271785"
I0909 10:44:56.100796 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.097702 16785 addons.go:69] Setting gcp-auth=true in profile "addons-271785"
I0909 10:44:56.097715 16785 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-271785"
I0909 10:44:56.097675 16785 addons.go:69] Setting ingress=true in profile "addons-271785"
I0909 10:44:56.099260 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.102107 16785 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0909 10:44:56.105048 16785 mustload.go:65] Loading cluster: addons-271785
I0909 10:44:56.105305 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.105491 16785 addons.go:234] Setting addon ingress=true in "addons-271785"
I0909 10:44:56.105561 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.106133 16785 config.go:182] Loaded profile config "addons-271785": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0909 10:44:56.107318 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.107566 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.107679 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.119337 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.135532 16785 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0909 10:44:56.137780 16785 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0909 10:44:56.137823 16785 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0909 10:44:56.137913 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.148174 16785 out.go:177] - Using image docker.io/registry:2.8.3
I0909 10:44:56.149442 16785 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0909 10:44:56.149572 16785 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0909 10:44:56.150053 16785 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0909 10:44:56.150516 16785 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0909 10:44:56.150535 16785 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0909 10:44:56.150599 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.151266 16785 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0909 10:44:56.151330 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0909 10:44:56.151432 16785 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0909 10:44:56.151442 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0909 10:44:56.151489 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.151716 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.155046 16785 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0909 10:44:56.156307 16785 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0909 10:44:56.156328 16785 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0909 10:44:56.156375 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.159008 16785 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0909 10:44:56.159975 16785 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0909 10:44:56.161063 16785 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0909 10:44:56.161067 16785 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0909 10:44:56.162356 16785 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0909 10:44:56.162444 16785 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0909 10:44:56.162567 16785 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0909 10:44:56.162595 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0909 10:44:56.162666 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.165086 16785 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0909 10:44:56.165144 16785 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0909 10:44:56.165261 16785 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0909 10:44:56.165277 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0909 10:44:56.165327 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.166303 16785 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0909 10:44:56.166322 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0909 10:44:56.166377 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.167296 16785 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0909 10:44:56.167362 16785 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0909 10:44:56.168406 16785 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0909 10:44:56.168427 16785 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0909 10:44:56.168442 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0909 10:44:56.168491 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.171309 16785 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-271785"
I0909 10:44:56.171350 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.171647 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.171864 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.173452 16785 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0909 10:44:56.174550 16785 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0909 10:44:56.175639 16785 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0909 10:44:56.176657 16785 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0909 10:44:56.176679 16785 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0909 10:44:56.176732 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.187969 16785 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0909 10:44:56.194320 16785 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0909 10:44:56.194342 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0909 10:44:56.194390 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.194514 16785 addons.go:234] Setting addon default-storageclass=true in "addons-271785"
I0909 10:44:56.194551 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:44:56.194979 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:44:56.196717 16785 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0909 10:44:56.197939 16785 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0909 10:44:56.197969 16785 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0909 10:44:56.198021 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.217135 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.219642 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.219632 16785 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0909 10:44:56.220451 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.222927 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.224431 16785 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0909 10:44:56.225548 16785 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0909 10:44:56.226842 16785 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0909 10:44:56.226859 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0909 10:44:56.226908 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.231541 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.232465 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.236275 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.244425 16785 out.go:177] - Using image docker.io/busybox:stable
I0909 10:44:56.247212 16785 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0909 10:44:56.248442 16785 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0909 10:44:56.248457 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0909 10:44:56.248497 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.248963 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.257239 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.261881 16785 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0909 10:44:56.261902 16785 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0909 10:44:56.261952 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:44:56.263778 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.265366 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.270457 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
W0909 10:44:56.272769 16785 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0909 10:44:56.272802 16785 retry.go:31] will retry after 193.188709ms: ssh: handshake failed: EOF
I0909 10:44:56.279432 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.294769 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.297572 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:44:56.464902 16785 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0909 10:44:56.465003 16785 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0909 10:44:56.569471 16785 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0909 10:44:56.569560 16785 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0909 10:44:56.650919 16785 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0909 10:44:56.650946 16785 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0909 10:44:56.658728 16785 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0909 10:44:56.658751 16785 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0909 10:44:56.664305 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0909 10:44:56.673877 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0909 10:44:56.772543 16785 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0909 10:44:56.772581 16785 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0909 10:44:56.853673 16785 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0909 10:44:56.853708 16785 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0909 10:44:56.863202 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0909 10:44:56.868435 16785 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0909 10:44:56.868463 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0909 10:44:56.873626 16785 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0909 10:44:56.873648 16785 ssh_runner.go:362] scp helm-tiller/helm-tiller-rbac.yaml --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0909 10:44:56.955532 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0909 10:44:56.972552 16785 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0909 10:44:56.972589 16785 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0909 10:44:57.049486 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0909 10:44:57.049650 16785 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0909 10:44:57.049699 16785 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0909 10:44:57.050762 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0909 10:44:57.062050 16785 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0909 10:44:57.062123 16785 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0909 10:44:57.064221 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0909 10:44:57.070389 16785 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0909 10:44:57.070457 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0909 10:44:57.071549 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0909 10:44:57.153719 16785 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0909 10:44:57.153807 16785 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0909 10:44:57.169339 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0909 10:44:57.264148 16785 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0909 10:44:57.264177 16785 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0909 10:44:57.352039 16785 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0909 10:44:57.352124 16785 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0909 10:44:57.369475 16785 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0909 10:44:57.369561 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0909 10:44:57.370644 16785 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0909 10:44:57.370727 16785 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0909 10:44:57.459488 16785 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0909 10:44:57.459575 16785 ssh_runner.go:362] scp helm-tiller/helm-tiller-svc.yaml --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0909 10:44:57.459864 16785 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0909 10:44:57.459913 16785 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0909 10:44:57.857162 16785 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0909 10:44:57.857266 16785 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0909 10:44:57.863080 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0909 10:44:57.964482 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0909 10:44:57.968810 16785 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0909 10:44:57.968888 16785 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0909 10:44:58.068368 16785 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0909 10:44:58.068461 16785 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0909 10:44:58.268715 16785 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0909 10:44:58.268767 16785 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0909 10:44:58.350129 16785 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0909 10:44:58.350187 16785 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0909 10:44:58.363269 16785 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0909 10:44:58.363297 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0909 10:44:58.451284 16785 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.986255838s)
I0909 10:44:58.452243 16785 node_ready.go:35] waiting up to 6m0s for node "addons-271785" to be "Ready" ...
I0909 10:44:58.452360 16785 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.987426079s)
I0909 10:44:58.452482 16785 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0909 10:44:58.458791 16785 node_ready.go:49] node "addons-271785" has status "Ready":"True"
I0909 10:44:58.458859 16785 node_ready.go:38] duration metric: took 6.47437ms for node "addons-271785" to be "Ready" ...
I0909 10:44:58.458884 16785 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0909 10:44:58.469292 16785 pod_ready.go:79] waiting up to 6m0s for pod "coredns-6f6b679f8f-dffc2" in "kube-system" namespace to be "Ready" ...
I0909 10:44:58.656111 16785 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0909 10:44:58.656199 16785 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0909 10:44:58.850324 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (2.185978948s)
I0909 10:44:58.853736 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0909 10:44:58.958762 16785 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-271785" context rescaled to 1 replicas
I0909 10:44:58.975251 16785 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0909 10:44:58.975312 16785 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0909 10:44:59.257094 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0909 10:44:59.767300 16785 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0909 10:44:59.767356 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0909 10:44:59.967959 16785 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0909 10:44:59.967989 16785 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0909 10:45:00.353016 16785 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0909 10:45:00.353044 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0909 10:45:00.354625 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0909 10:45:00.551818 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-dffc2" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:00.957500 16785 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0909 10:45:00.957533 16785 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0909 10:45:01.668817 16785 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0909 10:45:01.668847 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0909 10:45:02.067015 16785 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0909 10:45:02.067092 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0909 10:45:02.561886 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-dffc2" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:02.565848 16785 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0909 10:45:02.565929 16785 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0909 10:45:02.764624 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0909 10:45:03.257879 16785 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0909 10:45:03.258021 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:45:03.281844 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:45:03.949364 16785 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0909 10:45:04.168004 16785 addons.go:234] Setting addon gcp-auth=true in "addons-271785"
I0909 10:45:04.168075 16785 host.go:66] Checking if "addons-271785" exists ...
I0909 10:45:04.168621 16785 cli_runner.go:164] Run: docker container inspect addons-271785 --format={{.State.Status}}
I0909 10:45:04.187135 16785 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0909 10:45:04.187177 16785 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-271785
I0909 10:45:04.203022 16785 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19584-8635/.minikube/machines/addons-271785/id_rsa Username:docker}
I0909 10:45:05.054295 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-dffc2" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:07.151649 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-dffc2" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:08.058890 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.384979829s)
I0909 10:45:08.059119 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (11.103496825s)
I0909 10:45:08.059156 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (11.008345189s)
I0909 10:45:08.059185 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (11.009627596s)
I0909 10:45:08.059292 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (10.995007178s)
I0909 10:45:08.060134 16785 addons.go:475] Verifying addon ingress=true in "addons-271785"
I0909 10:45:08.059337 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (10.987734542s)
I0909 10:45:08.059372 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (10.889946368s)
I0909 10:45:08.060244 16785 addons.go:475] Verifying addon registry=true in "addons-271785"
I0909 10:45:08.059412 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (10.19624253s)
I0909 10:45:08.059450 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (10.094939165s)
I0909 10:45:08.059510 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (9.205742943s)
I0909 10:45:08.059638 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (8.802493605s)
I0909 10:45:08.060508 16785 addons.go:475] Verifying addon metrics-server=true in "addons-271785"
W0909 10:45:08.060525 16785 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0909 10:45:08.059711 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (7.705045657s)
I0909 10:45:08.060546 16785 retry.go:31] will retry after 276.723913ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0909 10:45:08.059746 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (11.195799082s)
I0909 10:45:08.062009 16785 out.go:177] * Verifying registry addon...
I0909 10:45:08.062029 16785 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-271785 service yakd-dashboard -n yakd-dashboard
I0909 10:45:08.062169 16785 out.go:177] * Verifying ingress addon...
I0909 10:45:08.064880 16785 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0909 10:45:08.064906 16785 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0909 10:45:08.069577 16785 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0909 10:45:08.069599 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:08.069977 16785 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0909 10:45:08.069992 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
W0909 10:45:08.149372 16785 out.go:270] ! Enabling 'default-storageclass' returned an error: running callbacks: [Error making standard the default storage class: Error while marking storage class local-path as non-default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
I0909 10:45:08.337782 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0909 10:45:08.570022 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:08.571124 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:09.070194 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:09.070531 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:09.500087 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-dffc2" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:09.553067 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (6.788391139s)
I0909 10:45:09.553109 16785 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.365949977s)
I0909 10:45:09.553104 16785 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-271785"
I0909 10:45:09.587309 16785 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0909 10:45:09.588405 16785 out.go:177] * Verifying csi-hostpath-driver addon...
I0909 10:45:09.591320 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:09.591908 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:09.593919 16785 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0909 10:45:09.594165 16785 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0909 10:45:09.596017 16785 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0909 10:45:09.596071 16785 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0909 10:45:09.598729 16785 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0909 10:45:09.598757 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:09.673389 16785 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0909 10:45:09.673424 16785 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0909 10:45:09.866459 16785 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0909 10:45:09.866485 16785 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0909 10:45:09.887782 16785 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0909 10:45:10.069658 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:10.071768 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:10.153043 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:10.475226 16785 pod_ready.go:98] pod "coredns-6f6b679f8f-dffc2" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-09 10:45:10 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-09 10:44:56 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-09 10:44:56 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-09 10:44:56 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-09 10:44:56 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.49.2 HostIPs:[{IP:192.168.49.2
}] PodIP:10.244.0.3 PodIPs:[{IP:10.244.0.3}] StartTime:2024-09-09 10:44:56 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-09-09 10:44:58 +0000 UTC,FinishedAt:2024-09-09 10:45:09 +0000 UTC,ContainerID:docker://c95c82b4e7e4c3b4dd355c99b8fcb5924f5cf81d7836e8fc817ddfc75475fdcc,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.1 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1 ContainerID:docker://c95c82b4e7e4c3b4dd355c99b8fcb5924f5cf81d7836e8fc817ddfc75475fdcc Started:0xc001aa4ac0 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0xc0008c5270} {Name:kube-api-access-4tpwt MountPath:/var/run/secrets/kubernetes.io/serviceaccount
ReadOnly:true RecursiveReadOnly:0xc0008c5280}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
I0909 10:45:10.475259 16785 pod_ready.go:82] duration metric: took 12.005887334s for pod "coredns-6f6b679f8f-dffc2" in "kube-system" namespace to be "Ready" ...
E0909 10:45:10.475274 16785 pod_ready.go:67] WaitExtra: waitPodCondition: pod "coredns-6f6b679f8f-dffc2" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-09 10:45:10 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-09 10:44:56 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-09 10:44:56 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-09 10:44:56 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-09 10:44:56 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.4
9.2 HostIPs:[{IP:192.168.49.2}] PodIP:10.244.0.3 PodIPs:[{IP:10.244.0.3}] StartTime:2024-09-09 10:44:56 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-09-09 10:44:58 +0000 UTC,FinishedAt:2024-09-09 10:45:09 +0000 UTC,ContainerID:docker://c95c82b4e7e4c3b4dd355c99b8fcb5924f5cf81d7836e8fc817ddfc75475fdcc,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.1 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1 ContainerID:docker://c95c82b4e7e4c3b4dd355c99b8fcb5924f5cf81d7836e8fc817ddfc75475fdcc Started:0xc001aa4ac0 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0xc0008c5270} {Name:kube-api-access-4tpwt MountPath:/var/run/secrets
/kubernetes.io/serviceaccount ReadOnly:true RecursiveReadOnly:0xc0008c5280}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
I0909 10:45:10.475283 16785 pod_ready.go:79] waiting up to 6m0s for pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace to be "Ready" ...
I0909 10:45:10.570120 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:10.570816 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:10.652947 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:10.872712 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.534881175s)
I0909 10:45:11.069593 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:11.070556 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:11.170447 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:11.255517 16785 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.367690263s)
I0909 10:45:11.257583 16785 addons.go:475] Verifying addon gcp-auth=true in "addons-271785"
I0909 10:45:11.259250 16785 out.go:177] * Verifying gcp-auth addon...
I0909 10:45:11.261237 16785 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0909 10:45:11.269048 16785 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0909 10:45:11.568216 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:11.569226 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:11.598062 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:12.068954 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:12.069304 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:12.098181 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:12.481144 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:12.569449 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:12.569772 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:12.597732 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:13.068546 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:13.069281 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:13.098615 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:13.569138 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:13.569576 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:13.598415 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:14.068907 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:14.069285 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:14.097672 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:14.568747 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:14.569058 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:14.677426 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:15.067250 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:15.067870 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:15.068159 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:15.097979 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:15.568461 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:15.568908 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:15.597471 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:16.068318 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:16.068868 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:16.097380 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:16.568872 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:16.569385 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:16.597998 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:17.068381 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:17.069040 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:17.097521 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:17.480319 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:17.568488 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:17.569057 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:17.597621 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:18.068319 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:18.068713 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:18.097677 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:18.568657 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:18.569184 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:18.598661 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:19.068899 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:19.069162 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:19.098098 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:19.480555 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:19.568315 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:19.568752 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:19.599198 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:20.068439 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:20.068916 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:20.097566 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:20.568421 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:20.569016 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:20.597468 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:21.069238 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:21.069462 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:21.098398 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:21.569484 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:21.570462 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:21.597994 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:21.981327 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:22.068879 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:22.069348 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:22.098390 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:22.568374 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:22.568826 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:22.599180 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:23.069048 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:23.069175 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:23.098405 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:23.569492 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:23.570007 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:23.598416 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:23.981356 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:24.068373 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:24.068890 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:24.098476 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:24.569101 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:24.569593 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:24.598757 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:25.068106 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:25.069113 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:25.098067 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:25.569023 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:25.569569 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:25.598522 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:26.140822 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:26.141092 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:26.141247 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:26.480628 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:26.569004 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:26.569075 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:26.598248 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:27.068346 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:27.068687 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:27.098428 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:27.568935 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:27.569196 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:27.599510 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:28.069081 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:28.069260 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:28.170449 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:28.567983 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:28.568397 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:28.598119 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:28.980886 16785 pod_ready.go:103] pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace has status "Ready":"False"
I0909 10:45:29.069007 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:29.069201 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:29.098397 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:29.568666 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:29.568925 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:29.597581 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:30.068463 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:30.068756 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:30.098225 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:30.568932 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:30.569134 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:30.597670 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:30.981655 16785 pod_ready.go:93] pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace has status "Ready":"True"
I0909 10:45:30.981681 16785 pod_ready.go:82] duration metric: took 20.506385633s for pod "coredns-6f6b679f8f-xc4cc" in "kube-system" namespace to be "Ready" ...
I0909 10:45:30.981699 16785 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-271785" in "kube-system" namespace to be "Ready" ...
I0909 10:45:30.986343 16785 pod_ready.go:93] pod "etcd-addons-271785" in "kube-system" namespace has status "Ready":"True"
I0909 10:45:30.986364 16785 pod_ready.go:82] duration metric: took 4.657734ms for pod "etcd-addons-271785" in "kube-system" namespace to be "Ready" ...
I0909 10:45:30.986375 16785 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-271785" in "kube-system" namespace to be "Ready" ...
I0909 10:45:30.990565 16785 pod_ready.go:93] pod "kube-apiserver-addons-271785" in "kube-system" namespace has status "Ready":"True"
I0909 10:45:30.990587 16785 pod_ready.go:82] duration metric: took 4.204614ms for pod "kube-apiserver-addons-271785" in "kube-system" namespace to be "Ready" ...
I0909 10:45:30.990597 16785 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-271785" in "kube-system" namespace to be "Ready" ...
I0909 10:45:30.994770 16785 pod_ready.go:93] pod "kube-controller-manager-addons-271785" in "kube-system" namespace has status "Ready":"True"
I0909 10:45:30.994787 16785 pod_ready.go:82] duration metric: took 4.184349ms for pod "kube-controller-manager-addons-271785" in "kube-system" namespace to be "Ready" ...
I0909 10:45:30.994795 16785 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-2qw8w" in "kube-system" namespace to be "Ready" ...
I0909 10:45:30.998855 16785 pod_ready.go:93] pod "kube-proxy-2qw8w" in "kube-system" namespace has status "Ready":"True"
I0909 10:45:30.998876 16785 pod_ready.go:82] duration metric: took 4.074243ms for pod "kube-proxy-2qw8w" in "kube-system" namespace to be "Ready" ...
I0909 10:45:30.998887 16785 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-271785" in "kube-system" namespace to be "Ready" ...
I0909 10:45:31.069763 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:31.070165 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:31.098482 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:31.379119 16785 pod_ready.go:93] pod "kube-scheduler-addons-271785" in "kube-system" namespace has status "Ready":"True"
I0909 10:45:31.379146 16785 pod_ready.go:82] duration metric: took 380.250233ms for pod "kube-scheduler-addons-271785" in "kube-system" namespace to be "Ready" ...
I0909 10:45:31.379159 16785 pod_ready.go:39] duration metric: took 32.92025059s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0909 10:45:31.379186 16785 api_server.go:52] waiting for apiserver process to appear ...
I0909 10:45:31.379244 16785 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0909 10:45:31.395676 16785 api_server.go:72] duration metric: took 35.298217523s to wait for apiserver process to appear ...
I0909 10:45:31.395708 16785 api_server.go:88] waiting for apiserver healthz status ...
I0909 10:45:31.395728 16785 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0909 10:45:31.399984 16785 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0909 10:45:31.400942 16785 api_server.go:141] control plane version: v1.31.0
I0909 10:45:31.400965 16785 api_server.go:131] duration metric: took 5.250142ms to wait for apiserver health ...
I0909 10:45:31.400975 16785 system_pods.go:43] waiting for kube-system pods to appear ...
I0909 10:45:31.568794 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:31.569391 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:31.585775 16785 system_pods.go:59] 18 kube-system pods found
I0909 10:45:31.585805 16785 system_pods.go:61] "coredns-6f6b679f8f-xc4cc" [4a149e0a-43a5-44e7-b91e-439700db0ec3] Running
I0909 10:45:31.585817 16785 system_pods.go:61] "csi-hostpath-attacher-0" [c667af28-28dd-41e6-86a9-6311794cfe78] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0909 10:45:31.585826 16785 system_pods.go:61] "csi-hostpath-resizer-0" [a3991825-0f50-4d74-8f14-07c871058034] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0909 10:45:31.585840 16785 system_pods.go:61] "csi-hostpathplugin-kn528" [9bf84cba-aca0-46f3-827f-73bd8b182cf1] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0909 10:45:31.585853 16785 system_pods.go:61] "etcd-addons-271785" [b71718eb-054a-4734-93f9-b77e30d49ec5] Running
I0909 10:45:31.585860 16785 system_pods.go:61] "kube-apiserver-addons-271785" [6c1744ba-36d7-4393-80e0-88725d0386ef] Running
I0909 10:45:31.585866 16785 system_pods.go:61] "kube-controller-manager-addons-271785" [6e95f0ee-dfdd-489d-9662-404dafe6b803] Running
I0909 10:45:31.585878 16785 system_pods.go:61] "kube-ingress-dns-minikube" [cadce7d8-dd7a-44de-a164-c2fba6ede595] Running
I0909 10:45:31.585883 16785 system_pods.go:61] "kube-proxy-2qw8w" [98ba3c1f-cf11-47aa-9e7d-393934752a66] Running
I0909 10:45:31.585890 16785 system_pods.go:61] "kube-scheduler-addons-271785" [8265d138-c7ec-44cc-907f-5fab14c8d119] Running
I0909 10:45:31.585901 16785 system_pods.go:61] "metrics-server-84c5f94fbc-jmhrl" [b5e8b788-3dd6-4f12-ad84-911eedfe943d] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0909 10:45:31.585909 16785 system_pods.go:61] "nvidia-device-plugin-daemonset-tdngv" [e84a8fea-2be6-41b4-a429-3b434f6fcb8a] Running
I0909 10:45:31.585915 16785 system_pods.go:61] "registry-6fb4cdfc84-g5pxq" [19051672-048d-4f1c-8814-35c5fa1de42e] Running
I0909 10:45:31.585926 16785 system_pods.go:61] "registry-proxy-dsb8t" [ad21aca5-affd-4e4b-9d2e-487316ad11de] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0909 10:45:31.585934 16785 system_pods.go:61] "snapshot-controller-56fcc65765-jktkk" [db496d37-be6c-49d8-9f2c-71c3766e673f] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0909 10:45:31.585945 16785 system_pods.go:61] "snapshot-controller-56fcc65765-sgwq2" [e148e085-17d3-492b-81a7-a24dc305ea28] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0909 10:45:31.585951 16785 system_pods.go:61] "storage-provisioner" [e0811adb-b379-4e54-86e2-91d05060bc58] Running
I0909 10:45:31.585962 16785 system_pods.go:61] "tiller-deploy-b48cc5f79-jsllm" [e269fd50-e1f6-4fe0-a330-8302e46d81af] Pending / Ready:ContainersNotReady (containers with unready status: [tiller]) / ContainersReady:ContainersNotReady (containers with unready status: [tiller])
I0909 10:45:31.585970 16785 system_pods.go:74] duration metric: took 184.988588ms to wait for pod list to return data ...
I0909 10:45:31.585982 16785 default_sa.go:34] waiting for default service account to be created ...
I0909 10:45:31.598130 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:31.778967 16785 default_sa.go:45] found service account: "default"
I0909 10:45:31.778991 16785 default_sa.go:55] duration metric: took 193.00035ms for default service account to be created ...
I0909 10:45:31.779002 16785 system_pods.go:116] waiting for k8s-apps to be running ...
I0909 10:45:31.986306 16785 system_pods.go:86] 18 kube-system pods found
I0909 10:45:31.986337 16785 system_pods.go:89] "coredns-6f6b679f8f-xc4cc" [4a149e0a-43a5-44e7-b91e-439700db0ec3] Running
I0909 10:45:31.986351 16785 system_pods.go:89] "csi-hostpath-attacher-0" [c667af28-28dd-41e6-86a9-6311794cfe78] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0909 10:45:31.986360 16785 system_pods.go:89] "csi-hostpath-resizer-0" [a3991825-0f50-4d74-8f14-07c871058034] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0909 10:45:31.986370 16785 system_pods.go:89] "csi-hostpathplugin-kn528" [9bf84cba-aca0-46f3-827f-73bd8b182cf1] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0909 10:45:31.986377 16785 system_pods.go:89] "etcd-addons-271785" [b71718eb-054a-4734-93f9-b77e30d49ec5] Running
I0909 10:45:31.986383 16785 system_pods.go:89] "kube-apiserver-addons-271785" [6c1744ba-36d7-4393-80e0-88725d0386ef] Running
I0909 10:45:31.986389 16785 system_pods.go:89] "kube-controller-manager-addons-271785" [6e95f0ee-dfdd-489d-9662-404dafe6b803] Running
I0909 10:45:31.986397 16785 system_pods.go:89] "kube-ingress-dns-minikube" [cadce7d8-dd7a-44de-a164-c2fba6ede595] Running
I0909 10:45:31.986402 16785 system_pods.go:89] "kube-proxy-2qw8w" [98ba3c1f-cf11-47aa-9e7d-393934752a66] Running
I0909 10:45:31.986412 16785 system_pods.go:89] "kube-scheduler-addons-271785" [8265d138-c7ec-44cc-907f-5fab14c8d119] Running
I0909 10:45:31.986420 16785 system_pods.go:89] "metrics-server-84c5f94fbc-jmhrl" [b5e8b788-3dd6-4f12-ad84-911eedfe943d] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0909 10:45:31.986431 16785 system_pods.go:89] "nvidia-device-plugin-daemonset-tdngv" [e84a8fea-2be6-41b4-a429-3b434f6fcb8a] Running
I0909 10:45:31.986437 16785 system_pods.go:89] "registry-6fb4cdfc84-g5pxq" [19051672-048d-4f1c-8814-35c5fa1de42e] Running
I0909 10:45:31.986450 16785 system_pods.go:89] "registry-proxy-dsb8t" [ad21aca5-affd-4e4b-9d2e-487316ad11de] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0909 10:45:31.986462 16785 system_pods.go:89] "snapshot-controller-56fcc65765-jktkk" [db496d37-be6c-49d8-9f2c-71c3766e673f] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0909 10:45:31.986474 16785 system_pods.go:89] "snapshot-controller-56fcc65765-sgwq2" [e148e085-17d3-492b-81a7-a24dc305ea28] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0909 10:45:31.986485 16785 system_pods.go:89] "storage-provisioner" [e0811adb-b379-4e54-86e2-91d05060bc58] Running
I0909 10:45:31.986493 16785 system_pods.go:89] "tiller-deploy-b48cc5f79-jsllm" [e269fd50-e1f6-4fe0-a330-8302e46d81af] Pending / Ready:ContainersNotReady (containers with unready status: [tiller]) / ContainersReady:ContainersNotReady (containers with unready status: [tiller])
I0909 10:45:31.986504 16785 system_pods.go:126] duration metric: took 207.495896ms to wait for k8s-apps to be running ...
I0909 10:45:31.986517 16785 system_svc.go:44] waiting for kubelet service to be running ....
I0909 10:45:31.986565 16785 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0909 10:45:32.000045 16785 system_svc.go:56] duration metric: took 13.513404ms WaitForService to wait for kubelet
I0909 10:45:32.000078 16785 kubeadm.go:582] duration metric: took 35.902621946s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0909 10:45:32.000102 16785 node_conditions.go:102] verifying NodePressure condition ...
I0909 10:45:32.068837 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:32.069122 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:32.097653 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:32.180108 16785 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0909 10:45:32.180137 16785 node_conditions.go:123] node cpu capacity is 8
I0909 10:45:32.180152 16785 node_conditions.go:105] duration metric: took 180.043957ms to run NodePressure ...
I0909 10:45:32.180166 16785 start.go:241] waiting for startup goroutines ...
I0909 10:45:32.569106 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:32.569212 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:32.598799 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:33.068861 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:33.069345 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:33.098542 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:33.568860 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:33.569216 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:33.597991 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:34.068196 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:34.069237 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:34.098415 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:34.570243 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0909 10:45:34.570429 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:34.598413 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:35.068293 16785 kapi.go:107] duration metric: took 27.003409174s to wait for kubernetes.io/minikube-addons=registry ...
I0909 10:45:35.068955 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:35.098143 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:35.569935 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:35.598094 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:36.069210 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:36.098102 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:36.568874 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:36.597882 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:37.069042 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:37.098907 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:37.570098 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:37.598444 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:38.070137 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:38.171353 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:38.569058 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:38.598485 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:39.069114 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:39.098333 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:39.568745 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:39.597755 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:40.069265 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:40.098530 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:40.569480 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:40.599113 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:41.069216 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:41.098320 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:41.570016 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:41.598337 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:42.070078 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:42.098298 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:42.569743 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:42.599051 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:43.068988 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:43.097654 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:43.570006 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:43.597823 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:44.068831 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:44.098641 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:44.569562 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:44.598300 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:45.069579 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:45.098795 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:45.570419 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:45.598665 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:46.070110 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:46.098378 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:46.569338 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:46.598265 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:47.069168 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:47.098544 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:47.569848 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:47.597792 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:48.069550 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:48.153266 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:48.570271 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:48.598401 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:49.069711 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:49.098524 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:49.569264 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:49.597890 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:50.068738 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:50.097412 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:50.568935 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:50.597730 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:51.069122 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:51.098362 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:51.569762 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:51.598217 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:52.070023 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:52.098193 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:52.569210 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:52.597814 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:53.069453 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:53.098075 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:53.568361 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:53.598364 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:54.069194 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:54.098292 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:54.570206 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:54.597987 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:55.068705 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:55.097479 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:55.570400 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:55.597558 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:56.069126 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:56.097693 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:56.570393 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:56.598485 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:57.069931 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:57.098501 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:57.569627 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:57.670843 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:58.070053 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:58.098905 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:58.570841 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:58.598753 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:59.070194 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:59.098269 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:45:59.568385 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:45:59.598462 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:00.070267 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:00.098490 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:00.569450 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:00.598257 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:01.069775 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:01.098503 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:01.569321 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:01.598101 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:02.069957 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:02.152948 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:02.570307 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:02.598418 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:03.069426 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:03.098554 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:03.568931 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:03.602718 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:04.070177 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:04.097749 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:04.569892 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:04.598278 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:05.068745 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:05.097420 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:05.570025 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:05.598022 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0909 10:46:06.069289 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:06.098115 16785 kapi.go:107] duration metric: took 56.504195187s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0909 10:46:06.569028 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:07.068940 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:07.569037 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:08.069164 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:08.568812 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:09.068529 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:09.569686 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:10.069462 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:10.569484 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:11.069753 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:11.569621 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:12.069938 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:12.569362 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:13.069282 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:13.569457 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:14.069557 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:14.568891 16785 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0909 10:46:15.068722 16785 kapi.go:107] duration metric: took 1m7.003811979s to wait for app.kubernetes.io/name=ingress-nginx ...
I0909 10:46:34.764715 16785 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0909 10:46:34.764737 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:35.264623 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:35.765063 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:36.265024 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:36.765326 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:37.263940 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:37.765056 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:38.264919 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:38.765380 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:39.263959 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:39.764645 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:40.264048 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:40.764003 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:41.264695 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:41.764524 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:42.264301 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:42.763917 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:43.264626 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:43.764531 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:44.264004 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:44.765013 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:45.264470 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:45.764832 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:46.264720 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:46.764608 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:47.264039 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:47.765979 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:48.264331 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:48.764510 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:49.264142 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:49.765107 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:50.264837 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:50.764997 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:51.263973 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:51.765001 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:52.264348 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:52.764270 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:53.263934 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:53.765079 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:54.264846 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:54.764668 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:55.264196 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:55.764306 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:56.264185 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:56.764023 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:57.264374 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:57.764296 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:58.263822 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:58.764319 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:59.263979 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:46:59.765125 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:00.263754 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:00.764896 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:01.264367 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:01.764250 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:02.263817 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:02.764161 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:03.263896 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:03.764414 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:04.263992 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:04.765211 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:05.264032 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:05.764823 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:06.264921 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:06.764588 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:07.264078 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:07.765598 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:08.264027 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:08.764867 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:09.264096 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:09.764942 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:10.264659 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:10.764505 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:11.264143 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:11.764099 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:12.263463 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:12.764418 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:13.264034 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:13.765355 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:14.263573 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:14.764865 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:15.264369 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:15.764402 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:16.264183 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:16.763843 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:17.263969 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:17.764862 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:18.264324 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:18.764032 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:19.264689 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:19.764552 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:20.264334 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:20.764136 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:21.264345 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:21.764503 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:22.264641 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:22.764208 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:23.263703 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:23.764721 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:24.264397 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:24.764617 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:25.263959 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:25.764109 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:26.263865 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:26.764111 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:27.263650 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:27.764212 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:28.264893 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:28.764254 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:29.263987 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:29.764960 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:30.264790 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:30.764531 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:31.264904 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:31.764459 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:32.264032 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:32.764975 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:33.263910 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:33.764315 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:34.263926 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:34.765149 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:35.263488 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:35.764531 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:36.264404 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:36.764129 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:37.264719 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:37.764143 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:38.263974 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:38.764639 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:39.264669 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:39.764482 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:40.264144 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:40.763818 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:41.263845 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:41.764343 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:42.264030 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:42.764875 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:43.264379 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:43.764356 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:44.264193 16785 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0909 10:47:44.764106 16785 kapi.go:107] duration metric: took 2m33.502869802s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0909 10:47:44.765684 16785 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-271785 cluster.
I0909 10:47:44.766841 16785 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0909 10:47:44.767966 16785 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0909 10:47:44.769336 16785 out.go:177] * Enabled addons: cloud-spanner, volcano, storage-provisioner, nvidia-device-plugin, ingress-dns, helm-tiller, metrics-server, inspektor-gadget, yakd, storage-provisioner-rancher, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0909 10:47:44.770607 16785 addons.go:510] duration metric: took 2m48.673107629s for enable addons: enabled=[cloud-spanner volcano storage-provisioner nvidia-device-plugin ingress-dns helm-tiller metrics-server inspektor-gadget yakd storage-provisioner-rancher volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0909 10:47:44.770654 16785 start.go:246] waiting for cluster config update ...
I0909 10:47:44.770678 16785 start.go:255] writing updated cluster config ...
I0909 10:47:44.770936 16785 ssh_runner.go:195] Run: rm -f paused
I0909 10:47:44.818524 16785 start.go:600] kubectl: 1.31.0, cluster: 1.31.0 (minor skew: 0)
I0909 10:47:44.820215 16785 out.go:177] * Done! kubectl is now configured to use "addons-271785" cluster and "default" namespace by default
==> Docker <==
Sep 09 10:57:18 addons-271785 dockerd[1341]: time="2024-09-09T10:57:18.266392114Z" level=info msg="ignoring event" container=0ed45814ba1d2db508cb0e68be7078b2b52bf43533472832aa41e6e63c82fa9e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:18 addons-271785 dockerd[1341]: time="2024-09-09T10:57:18.269424386Z" level=info msg="ignoring event" container=4a3262d776fe3850b5b316d80bba59e1cb87b600e4ad18bf17da4238c927daae module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:18 addons-271785 dockerd[1341]: time="2024-09-09T10:57:18.439076732Z" level=info msg="ignoring event" container=5594a069000a201b004e0f9223cc815dd07988c20dd3af407d5963a34c7f2f0e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:18 addons-271785 dockerd[1341]: time="2024-09-09T10:57:18.478857068Z" level=info msg="ignoring event" container=22ce97c7ff82343c85276c9d775dbbe640e8eb1157083d15cc24dc80db1f63a7 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:19 addons-271785 cri-dockerd[1606]: time="2024-09-09T10:57:19Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/1cf044f9074b2bc33e433dccd76cd0343d0e4ea9b824e48236f5ecc1bdfd3617/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local europe-west1-b.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 09 10:57:22 addons-271785 cri-dockerd[1606]: time="2024-09-09T10:57:22Z" level=info msg="Stop pulling image docker.io/nginx:alpine: Status: Downloaded newer image for nginx:alpine"
Sep 09 10:57:22 addons-271785 dockerd[1341]: time="2024-09-09T10:57:22.789176529Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 09 10:57:22 addons-271785 dockerd[1341]: time="2024-09-09T10:57:22.791165480Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 09 10:57:25 addons-271785 dockerd[1341]: time="2024-09-09T10:57:25.582067653Z" level=info msg="Container failed to exit within 30s of signal 15 - using the force" container=d2900ca675659c048969b5583fa937c074208babd6254c61f7fde5d2acf2e61a
Sep 09 10:57:25 addons-271785 dockerd[1341]: time="2024-09-09T10:57:25.603905680Z" level=info msg="ignoring event" container=d2900ca675659c048969b5583fa937c074208babd6254c61f7fde5d2acf2e61a module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:25 addons-271785 dockerd[1341]: time="2024-09-09T10:57:25.725879392Z" level=info msg="ignoring event" container=72bf5273b84495c140c234db2fedaad8d856b9063907f62e100f0a6113fb11da module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:29 addons-271785 cri-dockerd[1606]: time="2024-09-09T10:57:29Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/b6bb3e4018f8c08351b178e5ee0762195d5fc804bd9d53055557367216d9ddd4/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local europe-west1-b.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 09 10:57:30 addons-271785 dockerd[1341]: time="2024-09-09T10:57:30.079371025Z" level=info msg="ignoring event" container=acef1a4f7f0225f5c9023625762d4b94e4f2573b56e82404b54438395a899e5c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:30 addons-271785 dockerd[1341]: time="2024-09-09T10:57:30.122610369Z" level=info msg="ignoring event" container=ec9d5179df8546a830110aa12470f6fb2eaa3f556bbdef37921e965154bf3fc6 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:31 addons-271785 cri-dockerd[1606]: time="2024-09-09T10:57:31Z" level=info msg="Stop pulling image docker.io/kicbase/echo-server:1.0: Status: Downloaded newer image for kicbase/echo-server:1.0"
Sep 09 10:57:31 addons-271785 cri-dockerd[1606]: time="2024-09-09T10:57:31Z" level=error msg="error getting RW layer size for container ID 'acef1a4f7f0225f5c9023625762d4b94e4f2573b56e82404b54438395a899e5c': Error response from daemon: No such container: acef1a4f7f0225f5c9023625762d4b94e4f2573b56e82404b54438395a899e5c"
Sep 09 10:57:31 addons-271785 cri-dockerd[1606]: time="2024-09-09T10:57:31Z" level=error msg="Set backoffDuration to : 1m0s for container ID 'acef1a4f7f0225f5c9023625762d4b94e4f2573b56e82404b54438395a899e5c'"
Sep 09 10:57:33 addons-271785 dockerd[1341]: time="2024-09-09T10:57:33.874197751Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=339378f2559d5c1e7c29dbd4eacc9e35065c3a0af1a748743bfedb7040d5798f
Sep 09 10:57:33 addons-271785 dockerd[1341]: time="2024-09-09T10:57:33.938507690Z" level=info msg="ignoring event" container=339378f2559d5c1e7c29dbd4eacc9e35065c3a0af1a748743bfedb7040d5798f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:34 addons-271785 dockerd[1341]: time="2024-09-09T10:57:34.074672709Z" level=info msg="ignoring event" container=3e3df20fb697e4ff3329d775d38727b31f39058cf85f4976bf9a1a7ef424e755 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:38 addons-271785 dockerd[1341]: time="2024-09-09T10:57:38.089542468Z" level=info msg="ignoring event" container=4e0e0fb3807cb35ec5139095e8044e948fca346a498e444472c5387f8694746e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:38 addons-271785 dockerd[1341]: time="2024-09-09T10:57:38.673580269Z" level=info msg="ignoring event" container=eb3666654387c563813b8183daaf520231a6e980722f06dbffe97190cf3d17ec module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:38 addons-271785 dockerd[1341]: time="2024-09-09T10:57:38.759994151Z" level=info msg="ignoring event" container=c21bf18dbff2c4c0727e1617a3d4f6f6e7152448111e048ab49811a590c3001d module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:38 addons-271785 dockerd[1341]: time="2024-09-09T10:57:38.839685302Z" level=info msg="ignoring event" container=6500ac5b8a7f6fa13baf598dc8a351a8113f1d4704e6252ea1e316fcb5fe2044 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 09 10:57:38 addons-271785 dockerd[1341]: time="2024-09-09T10:57:38.904705557Z" level=info msg="ignoring event" container=406dcf66bc4785517829a439955bc27e1d3c2c66510a0f339f9f3d63454815ab module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
896bb364882b3 kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 8 seconds ago Running hello-world-app 0 b6bb3e4018f8c hello-world-app-55bf9c44b4-hrz4s
a8a5e1c2e8d3d nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf 17 seconds ago Running nginx 0 1cf044f9074b2 nginx
4099d2673ec9d a416a98b71e22 44 seconds ago Exited helper-pod 0 e8b268ca19228 helper-pod-delete-pvc-88f0cab2-ac8e-4b40-842d-f0e3d852d155
246ab1459c224 busybox@sha256:34b191d63fbc93e25e275bfccf1b5365664e5ac28f06d974e8d50090fbb49f41 48 seconds ago Exited busybox 0 0d0a53e91d710 test-local-path
289cac7194941 gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 dbaaf980f78e5 gcp-auth-89d5ffd79-62gbp
0947080312b3d registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited patch 0 d066a550f31f6 ingress-nginx-admission-patch-bw7kt
16c89e1623913 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 4b419f266e6c1 ingress-nginx-admission-create-xd2cr
2874786bcc709 6e38f40d628db 12 minutes ago Running storage-provisioner 0 a80172553100e storage-provisioner
226a97d5587a0 cbb01a7bd410d 12 minutes ago Running coredns 0 139674d0fde52 coredns-6f6b679f8f-xc4cc
05c87dd74bb34 ad83b2ca7b09e 12 minutes ago Running kube-proxy 0 e12e79d589e5f kube-proxy-2qw8w
7cb656dbc9404 1766f54c897f0 12 minutes ago Running kube-scheduler 0 877ce324fcfac kube-scheduler-addons-271785
529f303634a5d 045733566833c 12 minutes ago Running kube-controller-manager 0 a940bba3a8707 kube-controller-manager-addons-271785
acb1801eea7f4 2e96e5913fc06 12 minutes ago Running etcd 0 209cda1845a20 etcd-addons-271785
e3d6d9406589c 604f5db92eaa8 12 minutes ago Running kube-apiserver 0 fd98540cc3bfe kube-apiserver-addons-271785
==> coredns [226a97d5587a] <==
[INFO] 10.244.0.7:47898 - 19112 "A IN registry.kube-system.svc.cluster.local.svc.cluster.local. udp 74 false 512" NXDOMAIN qr,aa,rd 167 0.000101777s
[INFO] 10.244.0.7:49714 - 38829 "AAAA IN registry.kube-system.svc.cluster.local.cluster.local. udp 70 false 512" NXDOMAIN qr,aa,rd 163 0.000052929s
[INFO] 10.244.0.7:49714 - 40111 "A IN registry.kube-system.svc.cluster.local.cluster.local. udp 70 false 512" NXDOMAIN qr,aa,rd 163 0.000072918s
[INFO] 10.244.0.7:56501 - 27099 "A IN registry.kube-system.svc.cluster.local.europe-west1-b.c.k8s-minikube.internal. udp 95 false 512" NXDOMAIN qr,rd,ra 95 0.004472368s
[INFO] 10.244.0.7:56501 - 34271 "AAAA IN registry.kube-system.svc.cluster.local.europe-west1-b.c.k8s-minikube.internal. udp 95 false 512" NXDOMAIN qr,rd,ra 95 0.008068028s
[INFO] 10.244.0.7:51555 - 22310 "AAAA IN registry.kube-system.svc.cluster.local.c.k8s-minikube.internal. udp 80 false 512" NXDOMAIN qr,rd,ra 80 0.004923762s
[INFO] 10.244.0.7:51555 - 58393 "A IN registry.kube-system.svc.cluster.local.c.k8s-minikube.internal. udp 80 false 512" NXDOMAIN qr,rd,ra 80 0.00940149s
[INFO] 10.244.0.7:49232 - 527 "A IN registry.kube-system.svc.cluster.local.google.internal. udp 72 false 512" NXDOMAIN qr,rd,ra 72 0.004244518s
[INFO] 10.244.0.7:49232 - 25353 "AAAA IN registry.kube-system.svc.cluster.local.google.internal. udp 72 false 512" NXDOMAIN qr,rd,ra 72 0.014181837s
[INFO] 10.244.0.7:53833 - 30679 "AAAA IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 149 0.000068931s
[INFO] 10.244.0.7:53833 - 28881 "A IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 110 0.000109867s
[INFO] 10.244.0.26:47993 - 43472 "A IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000256472s
[INFO] 10.244.0.26:39436 - 11154 "AAAA IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000336047s
[INFO] 10.244.0.26:53627 - 63794 "A IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000118597s
[INFO] 10.244.0.26:53410 - 61496 "AAAA IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.00018251s
[INFO] 10.244.0.26:40694 - 30586 "A IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000103864s
[INFO] 10.244.0.26:60823 - 57652 "AAAA IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.00014825s
[INFO] 10.244.0.26:33363 - 25628 "AAAA IN storage.googleapis.com.europe-west1-b.c.k8s-minikube.internal. udp 90 false 1232" NXDOMAIN qr,rd,ra 79 0.006692325s
[INFO] 10.244.0.26:54746 - 45991 "A IN storage.googleapis.com.europe-west1-b.c.k8s-minikube.internal. udp 90 false 1232" NXDOMAIN qr,rd,ra 79 0.007446308s
[INFO] 10.244.0.26:40126 - 32577 "A IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.008136972s
[INFO] 10.244.0.26:50872 - 13500 "AAAA IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.009135357s
[INFO] 10.244.0.26:42237 - 41837 "AAAA IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.005136499s
[INFO] 10.244.0.26:37057 - 43691 "A IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.009384324s
[INFO] 10.244.0.26:50890 - 40458 "AAAA IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 140 0.001921982s
[INFO] 10.244.0.26:42884 - 53622 "A IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 116 0.002026719s
==> describe nodes <==
Name: addons-271785
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-271785
kubernetes.io/os=linux
minikube.k8s.io/commit=cf17d6b4040a54caaa170f92a048a513bb2a2b0d
minikube.k8s.io/name=addons-271785
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_09T10_44_51_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-271785
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Mon, 09 Sep 2024 10:44:48 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-271785
AcquireTime: <unset>
RenewTime: Mon, 09 Sep 2024 10:57:36 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Mon, 09 Sep 2024 10:57:27 +0000 Mon, 09 Sep 2024 10:44:47 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Mon, 09 Sep 2024 10:57:27 +0000 Mon, 09 Sep 2024 10:44:47 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Mon, 09 Sep 2024 10:57:27 +0000 Mon, 09 Sep 2024 10:44:47 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Mon, 09 Sep 2024 10:57:27 +0000 Mon, 09 Sep 2024 10:44:48 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-271785
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859320Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859320Ki
pods: 110
System Info:
Machine ID: e5c0832c58ff4e57a0457258493163ed
System UUID: 05d5e461-ab29-4115-89a4-f7fb02cb90b0
Boot ID: 51edb45c-9c14-46a0-b4bd-bdee90b8f8a3
Kernel Version: 5.15.0-1067-gcp
OS Image: Ubuntu 22.04.4 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.2.0
Kubelet Version: v1.31.0
Kube-Proxy Version:
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (11 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m14s
default hello-world-app-55bf9c44b4-hrz4s 0 (0%) 0 (0%) 0 (0%) 0 (0%) 10s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 21s
gcp-auth gcp-auth-89d5ffd79-62gbp 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
kube-system coredns-6f6b679f8f-xc4cc 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-271785 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-271785 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-271785 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-2qw8w 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-271785 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal NodeHasSufficientMemory 12m (x8 over 12m) kubelet Node addons-271785 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m (x7 over 12m) kubelet Node addons-271785 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m (x7 over 12m) kubelet Node addons-271785 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-271785 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-271785 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-271785 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-271785 event: Registered Node addons-271785 in Controller
==> dmesg <==
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 56 62 c7 cf 70 b8 08 06
[Sep 9 10:46] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff f6 37 25 22 c7 71 08 06
[ +0.333759] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 5e c9 5a cd 5d 31 08 06
[ +0.039911] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 46 b7 4f b0 f1 95 08 06
[ +10.766152] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 36 a3 16 d3 b6 1f 08 06
[ +1.026280] IPv4: martian source 10.244.0.1 from 10.244.0.23, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 0e a7 e0 5a 81 46 08 06
[Sep 9 10:47] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000015] ll header: 00000000: ff ff ff ff ff ff 7a 0b e5 40 23 d9 08 06
[ +0.109009] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 1a 44 87 05 2c 6b 08 06
[ +28.562989] IPv4: martian source 10.244.0.1 from 10.244.0.26, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 12 35 ce af 6a 6f 08 06
[ +0.000445] IPv4: martian source 10.244.0.26 from 10.244.0.2, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff ee 37 0d 65 3f 4f 08 06
[Sep 9 10:56] IPv4: martian source 10.244.0.1 from 10.244.0.29, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff 32 bb 6d 43 67 4c 08 06
[ +3.493563] IPv4: martian source 10.244.0.1 from 10.244.0.31, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 42 2e 1e a8 f0 8f 08 06
[Sep 9 10:57] IPv4: martian source 10.244.0.38 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 36 a3 16 d3 b6 1f 08 06
==> etcd [acb1801eea7f] <==
{"level":"info","ts":"2024-09-09T10:44:46.459982Z","caller":"embed/etcd.go:571","msg":"cmux::serve","address":"192.168.49.2:2380"}
{"level":"info","ts":"2024-09-09T10:44:47.191754Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc is starting a new election at term 1"}
{"level":"info","ts":"2024-09-09T10:44:47.191791Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became pre-candidate at term 1"}
{"level":"info","ts":"2024-09-09T10:44:47.191829Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgPreVoteResp from aec36adc501070cc at term 1"}
{"level":"info","ts":"2024-09-09T10:44:47.191854Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became candidate at term 2"}
{"level":"info","ts":"2024-09-09T10:44:47.191864Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-09T10:44:47.191871Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-09-09T10:44:47.191881Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-09T10:44:47.192818Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-271785 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-09-09T10:44:47.192899Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-09T10:44:47.192969Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-09T10:44:47.193009Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-09T10:44:47.193057Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-09T10:44:47.193082Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-09T10:44:47.193560Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-09T10:44:47.193640Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-09T10:44:47.193671Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-09T10:44:47.194180Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-09T10:44:47.194385Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-09T10:44:47.194891Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-09T10:44:47.195157Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-09-09T10:46:13.678895Z","caller":"traceutil/trace.go:171","msg":"trace[1524646620] transaction","detail":"{read_only:false; response_revision:1276; number_of_response:1; }","duration":"103.77411ms","start":"2024-09-09T10:46:13.575101Z","end":"2024-09-09T10:46:13.678875Z","steps":["trace[1524646620] 'process raft request' (duration: 45.238888ms)","trace[1524646620] 'compare' (duration: 58.442505ms)"],"step_count":2}
{"level":"info","ts":"2024-09-09T10:54:47.210330Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1905}
{"level":"info","ts":"2024-09-09T10:54:47.235294Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1905,"took":"24.37667ms","hash":4020332034,"current-db-size-bytes":9084928,"current-db-size":"9.1 MB","current-db-size-in-use-bytes":5066752,"current-db-size-in-use":"5.1 MB"}
{"level":"info","ts":"2024-09-09T10:54:47.235344Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":4020332034,"revision":1905,"compact-revision":-1}
==> gcp-auth [289cac719494] <==
2024/09/09 10:56:32 Ready to write response ...
2024/09/09 10:56:37 Ready to marshal response ...
2024/09/09 10:56:37 Ready to write response ...
2024/09/09 10:56:38 Ready to marshal response ...
2024/09/09 10:56:38 Ready to write response ...
2024/09/09 10:56:40 Ready to marshal response ...
2024/09/09 10:56:40 Ready to write response ...
2024/09/09 10:56:44 Ready to marshal response ...
2024/09/09 10:56:44 Ready to write response ...
2024/09/09 10:56:44 Ready to marshal response ...
2024/09/09 10:56:44 Ready to write response ...
2024/09/09 10:56:53 Ready to marshal response ...
2024/09/09 10:56:53 Ready to write response ...
2024/09/09 10:56:53 Ready to marshal response ...
2024/09/09 10:56:53 Ready to write response ...
2024/09/09 10:56:53 Ready to marshal response ...
2024/09/09 10:56:53 Ready to write response ...
2024/09/09 10:56:54 Ready to marshal response ...
2024/09/09 10:56:54 Ready to write response ...
2024/09/09 10:57:01 Ready to marshal response ...
2024/09/09 10:57:01 Ready to write response ...
2024/09/09 10:57:18 Ready to marshal response ...
2024/09/09 10:57:18 Ready to write response ...
2024/09/09 10:57:29 Ready to marshal response ...
2024/09/09 10:57:29 Ready to write response ...
==> kernel <==
10:57:39 up 39 min, 0 users, load average: 1.25, 0.56, 0.33
Linux addons-271785 5.15.0-1067-gcp #75~20.04.1-Ubuntu SMP Wed Aug 7 20:43:22 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.4 LTS"
==> kube-apiserver [e3d6d9406589] <==
W0909 10:48:18.068121 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
E0909 10:56:35.904225 1 conn.go:339] Error on socket receive: read tcp 192.168.49.2:8443->192.168.49.1:58092: use of closed network connection
E0909 10:56:39.468976 1 upgradeaware.go:427] Error proxying data from client to backend: read tcp 192.168.49.2:8443->10.244.0.31:39238: read: connection reset by peer
I0909 10:56:42.212038 1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
I0909 10:56:47.206485 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0909 10:56:53.145825 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.109.198.163"}
E0909 10:57:10.881533 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
I0909 10:57:17.126426 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
I0909 10:57:18.113733 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0909 10:57:18.113779 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0909 10:57:18.150298 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0909 10:57:18.150346 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0909 10:57:18.163632 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0909 10:57:18.163695 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0909 10:57:18.174606 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0909 10:57:18.175218 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0909 10:57:18.175251 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0909 10:57:18.181406 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0909 10:57:18.181439 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0909 10:57:18.711000 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0909 10:57:18.910522 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.97.112.151"}
W0909 10:57:19.175261 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
W0909 10:57:19.181984 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
W0909 10:57:19.253066 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
I0909 10:57:29.457531 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.105.227.107"}
==> kube-controller-manager [529f303634a5] <==
W0909 10:57:27.641632 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0909 10:57:27.641684 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0909 10:57:27.917658 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-271785"
W0909 10:57:28.668328 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0909 10:57:28.668373 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0909 10:57:29.269029 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="11.373584ms"
I0909 10:57:29.272496 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="3.417967ms"
I0909 10:57:29.272605 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="70.492µs"
I0909 10:57:29.278334 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="38.731µs"
I0909 10:57:30.852905 1 job_controller.go:568] "enqueueing job" logger="job-controller" key="ingress-nginx/ingress-nginx-admission-create" delay="0s"
I0909 10:57:30.853286 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="ingress-nginx/ingress-nginx-controller-bc57996ff" duration="5.58µs"
I0909 10:57:30.856776 1 job_controller.go:568] "enqueueing job" logger="job-controller" key="ingress-nginx/ingress-nginx-admission-patch" delay="0s"
I0909 10:57:32.424776 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="4.636277ms"
I0909 10:57:32.424871 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="56.757µs"
W0909 10:57:33.638582 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0909 10:57:33.638622 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0909 10:57:35.945943 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0909 10:57:35.945978 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0909 10:57:38.520723 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-6fb4cdfc84" duration="5.059µs"
W0909 10:57:38.863916 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0909 10:57:38.863951 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0909 10:57:39.418127 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0909 10:57:39.418170 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0909 10:57:39.580307 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0909 10:57:39.580348 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
==> kube-proxy [05c87dd74bb3] <==
I0909 10:44:58.267657 1 server_linux.go:66] "Using iptables proxy"
I0909 10:44:58.950692 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0909 10:44:58.950787 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0909 10:44:59.261857 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0909 10:44:59.261924 1 server_linux.go:169] "Using iptables Proxier"
I0909 10:44:59.266405 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0909 10:44:59.266824 1 server.go:483] "Version info" version="v1.31.0"
I0909 10:44:59.266844 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0909 10:44:59.269418 1 config.go:197] "Starting service config controller"
I0909 10:44:59.269434 1 shared_informer.go:313] Waiting for caches to sync for service config
I0909 10:44:59.269455 1 config.go:104] "Starting endpoint slice config controller"
I0909 10:44:59.269460 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0909 10:44:59.270014 1 config.go:326] "Starting node config controller"
I0909 10:44:59.270023 1 shared_informer.go:313] Waiting for caches to sync for node config
I0909 10:44:59.370626 1 shared_informer.go:320] Caches are synced for node config
I0909 10:44:59.370669 1 shared_informer.go:320] Caches are synced for service config
I0909 10:44:59.370714 1 shared_informer.go:320] Caches are synced for endpoint slice config
==> kube-scheduler [7cb656dbc940] <==
W0909 10:44:48.263539 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0909 10:44:48.264240 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0909 10:44:48.263385 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0909 10:44:48.264527 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0909 10:44:48.263329 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0909 10:44:48.264708 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0909 10:44:48.265364 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
W0909 10:44:48.265403 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0909 10:44:48.265412 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0909 10:44:48.265429 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0909 10:44:48.265439 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0909 10:44:48.265480 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0909 10:44:48.265528 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0909 10:44:48.265550 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0909 10:44:48.265631 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0909 10:44:48.265685 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0909 10:44:48.265819 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0909 10:44:48.265851 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0909 10:44:49.121541 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0909 10:44:49.121580 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0909 10:44:49.199053 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0909 10:44:49.199092 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0909 10:44:49.449607 1 reflector.go:561] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0909 10:44:49.449649 1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
I0909 10:44:52.659681 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 09 10:57:34 addons-271785 kubelet[2438]: I0909 10:57:34.446928 2438 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"339378f2559d5c1e7c29dbd4eacc9e35065c3a0af1a748743bfedb7040d5798f"} err="failed to get container status \"339378f2559d5c1e7c29dbd4eacc9e35065c3a0af1a748743bfedb7040d5798f\": rpc error: code = Unknown desc = Error response from daemon: No such container: 339378f2559d5c1e7c29dbd4eacc9e35065c3a0af1a748743bfedb7040d5798f"
Sep 09 10:57:34 addons-271785 kubelet[2438]: E0909 10:57:34.663626 2438 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="28571f1b-06fe-4073-b6ac-9fe7c2f086c3"
Sep 09 10:57:34 addons-271785 kubelet[2438]: I0909 10:57:34.669013 2438 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89539b6c-c702-464d-a8db-4259f5518958" path="/var/lib/kubelet/pods/89539b6c-c702-464d-a8db-4259f5518958/volumes"
Sep 09 10:57:36 addons-271785 kubelet[2438]: E0909 10:57:36.663454 2438 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox\\\"\"" pod="default/registry-test" podUID="7b2b4c60-0540-4fa9-9b2c-1af08e001fcb"
Sep 09 10:57:38 addons-271785 kubelet[2438]: I0909 10:57:38.302592 2438 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxz5l\" (UniqueName: \"kubernetes.io/projected/7b2b4c60-0540-4fa9-9b2c-1af08e001fcb-kube-api-access-nxz5l\") pod \"7b2b4c60-0540-4fa9-9b2c-1af08e001fcb\" (UID: \"7b2b4c60-0540-4fa9-9b2c-1af08e001fcb\") "
Sep 09 10:57:38 addons-271785 kubelet[2438]: I0909 10:57:38.302639 2438 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/7b2b4c60-0540-4fa9-9b2c-1af08e001fcb-gcp-creds\") pod \"7b2b4c60-0540-4fa9-9b2c-1af08e001fcb\" (UID: \"7b2b4c60-0540-4fa9-9b2c-1af08e001fcb\") "
Sep 09 10:57:38 addons-271785 kubelet[2438]: I0909 10:57:38.302708 2438 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b2b4c60-0540-4fa9-9b2c-1af08e001fcb-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "7b2b4c60-0540-4fa9-9b2c-1af08e001fcb" (UID: "7b2b4c60-0540-4fa9-9b2c-1af08e001fcb"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Sep 09 10:57:38 addons-271785 kubelet[2438]: I0909 10:57:38.304366 2438 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b2b4c60-0540-4fa9-9b2c-1af08e001fcb-kube-api-access-nxz5l" (OuterVolumeSpecName: "kube-api-access-nxz5l") pod "7b2b4c60-0540-4fa9-9b2c-1af08e001fcb" (UID: "7b2b4c60-0540-4fa9-9b2c-1af08e001fcb"). InnerVolumeSpecName "kube-api-access-nxz5l". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 09 10:57:38 addons-271785 kubelet[2438]: I0909 10:57:38.403579 2438 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-nxz5l\" (UniqueName: \"kubernetes.io/projected/7b2b4c60-0540-4fa9-9b2c-1af08e001fcb-kube-api-access-nxz5l\") on node \"addons-271785\" DevicePath \"\""
Sep 09 10:57:38 addons-271785 kubelet[2438]: I0909 10:57:38.403612 2438 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/7b2b4c60-0540-4fa9-9b2c-1af08e001fcb-gcp-creds\") on node \"addons-271785\" DevicePath \"\""
Sep 09 10:57:38 addons-271785 kubelet[2438]: I0909 10:57:38.671782 2438 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b2b4c60-0540-4fa9-9b2c-1af08e001fcb" path="/var/lib/kubelet/pods/7b2b4c60-0540-4fa9-9b2c-1af08e001fcb/volumes"
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.051486 2438 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcwjc\" (UniqueName: \"kubernetes.io/projected/19051672-048d-4f1c-8814-35c5fa1de42e-kube-api-access-hcwjc\") pod \"19051672-048d-4f1c-8814-35c5fa1de42e\" (UID: \"19051672-048d-4f1c-8814-35c5fa1de42e\") "
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.051549 2438 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skz2j\" (UniqueName: \"kubernetes.io/projected/ad21aca5-affd-4e4b-9d2e-487316ad11de-kube-api-access-skz2j\") pod \"ad21aca5-affd-4e4b-9d2e-487316ad11de\" (UID: \"ad21aca5-affd-4e4b-9d2e-487316ad11de\") "
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.053618 2438 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19051672-048d-4f1c-8814-35c5fa1de42e-kube-api-access-hcwjc" (OuterVolumeSpecName: "kube-api-access-hcwjc") pod "19051672-048d-4f1c-8814-35c5fa1de42e" (UID: "19051672-048d-4f1c-8814-35c5fa1de42e"). InnerVolumeSpecName "kube-api-access-hcwjc". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.053732 2438 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad21aca5-affd-4e4b-9d2e-487316ad11de-kube-api-access-skz2j" (OuterVolumeSpecName: "kube-api-access-skz2j") pod "ad21aca5-affd-4e4b-9d2e-487316ad11de" (UID: "ad21aca5-affd-4e4b-9d2e-487316ad11de"). InnerVolumeSpecName "kube-api-access-skz2j". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.152109 2438 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-skz2j\" (UniqueName: \"kubernetes.io/projected/ad21aca5-affd-4e4b-9d2e-487316ad11de-kube-api-access-skz2j\") on node \"addons-271785\" DevicePath \"\""
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.152147 2438 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-hcwjc\" (UniqueName: \"kubernetes.io/projected/19051672-048d-4f1c-8814-35c5fa1de42e-kube-api-access-hcwjc\") on node \"addons-271785\" DevicePath \"\""
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.493816 2438 scope.go:117] "RemoveContainer" containerID="c21bf18dbff2c4c0727e1617a3d4f6f6e7152448111e048ab49811a590c3001d"
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.508637 2438 scope.go:117] "RemoveContainer" containerID="c21bf18dbff2c4c0727e1617a3d4f6f6e7152448111e048ab49811a590c3001d"
Sep 09 10:57:39 addons-271785 kubelet[2438]: E0909 10:57:39.509423 2438 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: c21bf18dbff2c4c0727e1617a3d4f6f6e7152448111e048ab49811a590c3001d" containerID="c21bf18dbff2c4c0727e1617a3d4f6f6e7152448111e048ab49811a590c3001d"
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.509478 2438 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"c21bf18dbff2c4c0727e1617a3d4f6f6e7152448111e048ab49811a590c3001d"} err="failed to get container status \"c21bf18dbff2c4c0727e1617a3d4f6f6e7152448111e048ab49811a590c3001d\": rpc error: code = Unknown desc = Error response from daemon: No such container: c21bf18dbff2c4c0727e1617a3d4f6f6e7152448111e048ab49811a590c3001d"
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.509506 2438 scope.go:117] "RemoveContainer" containerID="eb3666654387c563813b8183daaf520231a6e980722f06dbffe97190cf3d17ec"
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.525696 2438 scope.go:117] "RemoveContainer" containerID="eb3666654387c563813b8183daaf520231a6e980722f06dbffe97190cf3d17ec"
Sep 09 10:57:39 addons-271785 kubelet[2438]: E0909 10:57:39.526388 2438 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: eb3666654387c563813b8183daaf520231a6e980722f06dbffe97190cf3d17ec" containerID="eb3666654387c563813b8183daaf520231a6e980722f06dbffe97190cf3d17ec"
Sep 09 10:57:39 addons-271785 kubelet[2438]: I0909 10:57:39.526422 2438 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"eb3666654387c563813b8183daaf520231a6e980722f06dbffe97190cf3d17ec"} err="failed to get container status \"eb3666654387c563813b8183daaf520231a6e980722f06dbffe97190cf3d17ec\": rpc error: code = Unknown desc = Error response from daemon: No such container: eb3666654387c563813b8183daaf520231a6e980722f06dbffe97190cf3d17ec"
==> storage-provisioner [2874786bcc70] <==
I0909 10:45:04.860372 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0909 10:45:04.871764 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0909 10:45:04.871806 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0909 10:45:04.956675 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0909 10:45:04.957143 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-271785_a967797c-77a3-4845-9911-406176895cc0!
I0909 10:45:04.957137 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"82b53c78-dea4-46b6-be27-0396c43ccd5d", APIVersion:"v1", ResourceVersion:"673", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-271785_a967797c-77a3-4845-9911-406176895cc0 became leader
I0909 10:45:05.058110 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-271785_a967797c-77a3-4845-9911-406176895cc0!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-271785 -n addons-271785
helpers_test.go:261: (dbg) Run: kubectl --context addons-271785 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-271785 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-271785 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-271785/192.168.49.2
Start Time: Mon, 09 Sep 2024 10:48:25 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.28
IPs:
IP: 10.244.0.28
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-pr9fv (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-pr9fv:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m15s default-scheduler Successfully assigned default/busybox to addons-271785
Normal Pulling 7m44s (x4 over 9m14s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m44s (x4 over 9m14s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m44s (x4 over 9m14s) kubelet Error: ErrImagePull
Warning Failed 7m30s (x6 over 9m14s) kubelet Error: ImagePullBackOff
Normal BackOff 4m11s (x20 over 9m14s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (72.57s)