=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:332: registry stabilized in 2.440974ms
addons_test.go:334: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-xnftt" [87171e43-6b56-423a-ac20-6b46a3583197] Running
addons_test.go:334: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.002463391s
addons_test.go:337: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-9ztsk" [de43c7a6-1992-4444-969d-d41949e06cdb] Running
addons_test.go:337: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.002763958s
addons_test.go:342: (dbg) Run: kubectl --context addons-163060 delete po -l run=registry-test --now
addons_test.go:347: (dbg) Run: kubectl --context addons-163060 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:347: (dbg) Non-zero exit: kubectl --context addons-163060 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.077642698s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:349: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-163060 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:353: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:361: (dbg) Run: out/minikube-linux-amd64 -p addons-163060 ip
addons_test.go:390: (dbg) Run: out/minikube-linux-amd64 -p addons-163060 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-163060
helpers_test.go:235: (dbg) docker inspect addons-163060:
-- stdout --
[
{
"Id": "7a802f428c5e52f6745f399055a29e1c0f2dbe4f3db2c58b1b9fecaed240bff3",
"Created": "2024-09-17T16:56:27.16599634Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 20900,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-17T16:56:27.297240995Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:bb3bcbaabeeeadbf6b43ae7d1d07e504b3c8a94ec024df89bcb237eba4f5e9b3",
"ResolvConfPath": "/var/lib/docker/containers/7a802f428c5e52f6745f399055a29e1c0f2dbe4f3db2c58b1b9fecaed240bff3/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/7a802f428c5e52f6745f399055a29e1c0f2dbe4f3db2c58b1b9fecaed240bff3/hostname",
"HostsPath": "/var/lib/docker/containers/7a802f428c5e52f6745f399055a29e1c0f2dbe4f3db2c58b1b9fecaed240bff3/hosts",
"LogPath": "/var/lib/docker/containers/7a802f428c5e52f6745f399055a29e1c0f2dbe4f3db2c58b1b9fecaed240bff3/7a802f428c5e52f6745f399055a29e1c0f2dbe4f3db2c58b1b9fecaed240bff3-json.log",
"Name": "/addons-163060",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-163060:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-163060",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/b64b7857ceff83d215ea9bfa6f3ef683bfe63f9513851f53c908ce23e4bdb801-init/diff:/var/lib/docker/overlay2/03f685b8c3eedc410fe49fd5865e32dca92633e19bab382ce7cf454aa3c4e4e2/diff",
"MergedDir": "/var/lib/docker/overlay2/b64b7857ceff83d215ea9bfa6f3ef683bfe63f9513851f53c908ce23e4bdb801/merged",
"UpperDir": "/var/lib/docker/overlay2/b64b7857ceff83d215ea9bfa6f3ef683bfe63f9513851f53c908ce23e4bdb801/diff",
"WorkDir": "/var/lib/docker/overlay2/b64b7857ceff83d215ea9bfa6f3ef683bfe63f9513851f53c908ce23e4bdb801/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-163060",
"Source": "/var/lib/docker/volumes/addons-163060/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-163060",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-163060",
"name.minikube.sigs.k8s.io": "addons-163060",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "1baaaa9a9aefb5e6ad0de626340cf38f2a1dbaffaa513a1fa3fbe0b65e3c2f1c",
"SandboxKey": "/var/run/docker/netns/1baaaa9a9aef",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-163060": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "e13fb16c4676d315fab78c48ec3fa5ecd124d207fc37bae5ebb0dd7c50aa3999",
"EndpointID": "940717b186675aeb8ea59b0912119d652d86a69479c103cee94f6f00614b5331",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-163060",
"7a802f428c5e"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-163060 -n addons-163060
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-163060 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-docker-967799 | download-docker-967799 | jenkins | v1.34.0 | 17 Sep 24 16:56 UTC | 17 Sep 24 16:56 UTC |
| start | --download-only -p | binary-mirror-225195 | jenkins | v1.34.0 | 17 Sep 24 16:56 UTC | |
| | binary-mirror-225195 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:45015 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-225195 | binary-mirror-225195 | jenkins | v1.34.0 | 17 Sep 24 16:56 UTC | 17 Sep 24 16:56 UTC |
| addons | enable dashboard -p | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 16:56 UTC | |
| | addons-163060 | | | | | |
| addons | disable dashboard -p | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 16:56 UTC | |
| | addons-163060 | | | | | |
| start | -p addons-163060 --wait=true | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 16:56 UTC | 17 Sep 24 16:59 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | addons-163060 addons disable | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:00 UTC | 17 Sep 24 17:00 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:08 UTC | 17 Sep 24 17:08 UTC |
| | addons-163060 | | | | | |
| addons | addons-163060 addons | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:08 UTC | 17 Sep 24 17:08 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-163060 addons disable | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:08 UTC | 17 Sep 24 17:08 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:08 UTC | 17 Sep 24 17:08 UTC |
| | -p addons-163060 | | | | | |
| addons | disable cloud-spanner -p | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:08 UTC | 17 Sep 24 17:08 UTC |
| | addons-163060 | | | | | |
| addons | enable headlamp | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:08 UTC | 17 Sep 24 17:08 UTC |
| | -p addons-163060 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-163060 ssh cat | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:08 UTC | 17 Sep 24 17:08 UTC |
| | /opt/local-path-provisioner/pvc-6b40e24e-ff27-49e1-a0af-4a3320a2542e_default_test-pvc/file1 | | | | | |
| addons | addons-163060 addons disable | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:08 UTC | 17 Sep 24 17:09 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-163060 addons disable | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:08 UTC | 17 Sep 24 17:08 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ssh | addons-163060 ssh curl -s | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:09 UTC | 17 Sep 24 17:09 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-163060 ip | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:09 UTC | 17 Sep 24 17:09 UTC |
| addons | addons-163060 addons disable | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:09 UTC | 17 Sep 24 17:09 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-163060 addons disable | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:09 UTC | 17 Sep 24 17:09 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| addons | addons-163060 addons | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:09 UTC | 17 Sep 24 17:09 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-163060 addons | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:09 UTC | 17 Sep 24 17:09 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ip | addons-163060 ip | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:09 UTC | 17 Sep 24 17:09 UTC |
| addons | addons-163060 addons disable | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:09 UTC | 17 Sep 24 17:09 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-163060 addons disable | addons-163060 | jenkins | v1.34.0 | 17 Sep 24 17:09 UTC | 17 Sep 24 17:09 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/17 16:56:05
Running on machine: ubuntu-20-agent-10
Binary: Built with gc go1.23.0 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0917 16:56:05.429771 20150 out.go:345] Setting OutFile to fd 1 ...
I0917 16:56:05.429871 20150 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0917 16:56:05.429882 20150 out.go:358] Setting ErrFile to fd 2...
I0917 16:56:05.429889 20150 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0917 16:56:05.430090 20150 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19662-12004/.minikube/bin
I0917 16:56:05.430715 20150 out.go:352] Setting JSON to false
I0917 16:56:05.431645 20150 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-10","uptime":2306,"bootTime":1726589859,"procs":170,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1069-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0917 16:56:05.431741 20150 start.go:139] virtualization: kvm guest
I0917 16:56:05.433780 20150 out.go:177] * [addons-163060] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0917 16:56:05.435191 20150 notify.go:220] Checking for updates...
I0917 16:56:05.435203 20150 out.go:177] - MINIKUBE_LOCATION=19662
I0917 16:56:05.436538 20150 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0917 16:56:05.437862 20150 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19662-12004/kubeconfig
I0917 16:56:05.439080 20150 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19662-12004/.minikube
I0917 16:56:05.440324 20150 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0917 16:56:05.441547 20150 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0917 16:56:05.442990 20150 driver.go:394] Setting default libvirt URI to qemu:///system
I0917 16:56:05.463903 20150 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
I0917 16:56:05.463988 20150 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0917 16:56:05.508885 20150 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-17 16:56:05.500245979 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-10 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0917 16:56:05.508992 20150 docker.go:318] overlay module found
I0917 16:56:05.510795 20150 out.go:177] * Using the docker driver based on user configuration
I0917 16:56:05.511849 20150 start.go:297] selected driver: docker
I0917 16:56:05.511863 20150 start.go:901] validating driver "docker" against <nil>
I0917 16:56:05.511876 20150 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0917 16:56:05.512818 20150 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0917 16:56:05.554582 20150 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-17 16:56:05.546303107 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1069-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647943680 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-10 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0917 16:56:05.554745 20150 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0917 16:56:05.554985 20150 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0917 16:56:05.556575 20150 out.go:177] * Using Docker driver with root privileges
I0917 16:56:05.557944 20150 cni.go:84] Creating CNI manager for ""
I0917 16:56:05.557994 20150 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0917 16:56:05.558005 20150 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0917 16:56:05.558058 20150 start.go:340] cluster config:
{Name:addons-163060 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-163060 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0917 16:56:05.559264 20150 out.go:177] * Starting "addons-163060" primary control-plane node in "addons-163060" cluster
I0917 16:56:05.560282 20150 cache.go:121] Beginning downloading kic base image for docker with docker
I0917 16:56:05.561402 20150 out.go:177] * Pulling base image v0.0.45-1726589491-19662 ...
I0917 16:56:05.562546 20150 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0917 16:56:05.562571 20150 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19662-12004/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4
I0917 16:56:05.562588 20150 cache.go:56] Caching tarball of preloaded images
I0917 16:56:05.562636 20150 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 in local docker daemon
I0917 16:56:05.562670 20150 preload.go:172] Found /home/jenkins/minikube-integration/19662-12004/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0917 16:56:05.562681 20150 cache.go:59] Finished verifying existence of preloaded tar for v1.31.1 on docker
I0917 16:56:05.563063 20150 profile.go:143] Saving config to /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/config.json ...
I0917 16:56:05.563089 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/config.json: {Name:mkf8815af0780b232816101da29cb7accc3725b1 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:05.578629 20150 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 to local cache
I0917 16:56:05.578739 20150 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 in local cache directory
I0917 16:56:05.578761 20150 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 in local cache directory, skipping pull
I0917 16:56:05.578768 20150 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 exists in cache, skipping pull
I0917 16:56:05.578780 20150 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 as a tarball
I0917 16:56:05.578789 20150 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 from local cache
I0917 16:56:17.946938 20150 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 from cached tarball
I0917 16:56:17.946992 20150 cache.go:194] Successfully downloaded all kic artifacts
I0917 16:56:17.947029 20150 start.go:360] acquireMachinesLock for addons-163060: {Name:mk5177c6da83c393abee9e5f56591d5371bba180 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0917 16:56:17.947135 20150 start.go:364] duration metric: took 83.16µs to acquireMachinesLock for "addons-163060"
I0917 16:56:17.947166 20150 start.go:93] Provisioning new machine with config: &{Name:addons-163060 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-163060 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0917 16:56:17.947249 20150 start.go:125] createHost starting for "" (driver="docker")
I0917 16:56:17.950307 20150 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0917 16:56:17.950564 20150 start.go:159] libmachine.API.Create for "addons-163060" (driver="docker")
I0917 16:56:17.950622 20150 client.go:168] LocalClient.Create starting
I0917 16:56:17.950722 20150 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19662-12004/.minikube/certs/ca.pem
I0917 16:56:18.222813 20150 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19662-12004/.minikube/certs/cert.pem
I0917 16:56:18.282861 20150 cli_runner.go:164] Run: docker network inspect addons-163060 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0917 16:56:18.298058 20150 cli_runner.go:211] docker network inspect addons-163060 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0917 16:56:18.298118 20150 network_create.go:284] running [docker network inspect addons-163060] to gather additional debugging logs...
I0917 16:56:18.298137 20150 cli_runner.go:164] Run: docker network inspect addons-163060
W0917 16:56:18.312759 20150 cli_runner.go:211] docker network inspect addons-163060 returned with exit code 1
I0917 16:56:18.312790 20150 network_create.go:287] error running [docker network inspect addons-163060]: docker network inspect addons-163060: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-163060 not found
I0917 16:56:18.312803 20150 network_create.go:289] output of [docker network inspect addons-163060]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-163060 not found
** /stderr **
I0917 16:56:18.312905 20150 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0917 16:56:18.327849 20150 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001358570}
I0917 16:56:18.327892 20150 network_create.go:124] attempt to create docker network addons-163060 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0917 16:56:18.327932 20150 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-163060 addons-163060
I0917 16:56:18.385644 20150 network_create.go:108] docker network addons-163060 192.168.49.0/24 created
I0917 16:56:18.385678 20150 kic.go:121] calculated static IP "192.168.49.2" for the "addons-163060" container
I0917 16:56:18.385751 20150 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0917 16:56:18.399845 20150 cli_runner.go:164] Run: docker volume create addons-163060 --label name.minikube.sigs.k8s.io=addons-163060 --label created_by.minikube.sigs.k8s.io=true
I0917 16:56:18.416457 20150 oci.go:103] Successfully created a docker volume addons-163060
I0917 16:56:18.416541 20150 cli_runner.go:164] Run: docker run --rm --name addons-163060-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-163060 --entrypoint /usr/bin/test -v addons-163060:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -d /var/lib
I0917 16:56:23.304261 20150 cli_runner.go:217] Completed: docker run --rm --name addons-163060-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-163060 --entrypoint /usr/bin/test -v addons-163060:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -d /var/lib: (4.887680591s)
I0917 16:56:23.304302 20150 oci.go:107] Successfully prepared a docker volume addons-163060
I0917 16:56:23.304322 20150 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0917 16:56:23.304347 20150 kic.go:194] Starting extracting preloaded images to volume ...
I0917 16:56:23.304415 20150 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19662-12004/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-163060:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -I lz4 -xf /preloaded.tar -C /extractDir
I0917 16:56:27.107561 20150 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19662-12004/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-163060:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 -I lz4 -xf /preloaded.tar -C /extractDir: (3.80310416s)
I0917 16:56:27.107590 20150 kic.go:203] duration metric: took 3.803241274s to extract preloaded images to volume ...
W0917 16:56:27.107695 20150 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0917 16:56:27.107777 20150 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0917 16:56:27.152206 20150 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-163060 --name addons-163060 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-163060 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-163060 --network addons-163060 --ip 192.168.49.2 --volume addons-163060:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4
I0917 16:56:27.452182 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Running}}
I0917 16:56:27.471401 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:27.489466 20150 cli_runner.go:164] Run: docker exec addons-163060 stat /var/lib/dpkg/alternatives/iptables
I0917 16:56:27.528788 20150 oci.go:144] the created container "addons-163060" has a running status.
I0917 16:56:27.528821 20150 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa...
I0917 16:56:27.851877 20150 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0917 16:56:27.873995 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:27.898655 20150 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0917 16:56:27.898677 20150 kic_runner.go:114] Args: [docker exec --privileged addons-163060 chown docker:docker /home/docker/.ssh/authorized_keys]
I0917 16:56:27.957061 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:27.974688 20150 machine.go:93] provisionDockerMachine start ...
I0917 16:56:27.974767 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:27.992628 20150 main.go:141] libmachine: Using SSH client type: native
I0917 16:56:27.992891 20150 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 16:56:27.992912 20150 main.go:141] libmachine: About to run SSH command:
hostname
I0917 16:56:28.130251 20150 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-163060
I0917 16:56:28.130278 20150 ubuntu.go:169] provisioning hostname "addons-163060"
I0917 16:56:28.130345 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:28.147498 20150 main.go:141] libmachine: Using SSH client type: native
I0917 16:56:28.147703 20150 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 16:56:28.147720 20150 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-163060 && echo "addons-163060" | sudo tee /etc/hostname
I0917 16:56:28.288715 20150 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-163060
I0917 16:56:28.288790 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:28.304385 20150 main.go:141] libmachine: Using SSH client type: native
I0917 16:56:28.304592 20150 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 16:56:28.304612 20150 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-163060' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-163060/g' /etc/hosts;
else
echo '127.0.1.1 addons-163060' | sudo tee -a /etc/hosts;
fi
fi
I0917 16:56:28.434664 20150 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0917 16:56:28.434690 20150 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19662-12004/.minikube CaCertPath:/home/jenkins/minikube-integration/19662-12004/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19662-12004/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19662-12004/.minikube}
I0917 16:56:28.434730 20150 ubuntu.go:177] setting up certificates
I0917 16:56:28.434744 20150 provision.go:84] configureAuth start
I0917 16:56:28.434795 20150 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-163060
I0917 16:56:28.449572 20150 provision.go:143] copyHostCerts
I0917 16:56:28.449660 20150 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19662-12004/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19662-12004/.minikube/ca.pem (1082 bytes)
I0917 16:56:28.449778 20150 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19662-12004/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19662-12004/.minikube/cert.pem (1123 bytes)
I0917 16:56:28.449854 20150 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19662-12004/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19662-12004/.minikube/key.pem (1679 bytes)
I0917 16:56:28.449923 20150 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19662-12004/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19662-12004/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19662-12004/.minikube/certs/ca-key.pem org=jenkins.addons-163060 san=[127.0.0.1 192.168.49.2 addons-163060 localhost minikube]
I0917 16:56:28.742193 20150 provision.go:177] copyRemoteCerts
I0917 16:56:28.742249 20150 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0917 16:56:28.742283 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:28.758516 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:28.851213 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I0917 16:56:28.871585 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0917 16:56:28.891338 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0917 16:56:28.911541 20150 provision.go:87] duration metric: took 476.784156ms to configureAuth
I0917 16:56:28.911571 20150 ubuntu.go:193] setting minikube options for container-runtime
I0917 16:56:28.911740 20150 config.go:182] Loaded profile config "addons-163060": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0917 16:56:28.911796 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:28.927724 20150 main.go:141] libmachine: Using SSH client type: native
I0917 16:56:28.927917 20150 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 16:56:28.927936 20150 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0917 16:56:29.059042 20150 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0917 16:56:29.059071 20150 ubuntu.go:71] root file system type: overlay
I0917 16:56:29.059199 20150 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0917 16:56:29.059269 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:29.075003 20150 main.go:141] libmachine: Using SSH client type: native
I0917 16:56:29.075168 20150 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 16:56:29.075223 20150 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0917 16:56:29.217003 20150 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0917 16:56:29.217079 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:29.234540 20150 main.go:141] libmachine: Using SSH client type: native
I0917 16:56:29.234744 20150 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0917 16:56:29.234768 20150 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0917 16:56:29.890754 20150 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-09-06 12:06:41.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-17 16:56:29.211337128 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0917 16:56:29.890802 20150 machine.go:96] duration metric: took 1.916092594s to provisionDockerMachine
I0917 16:56:29.890816 20150 client.go:171] duration metric: took 11.940183267s to LocalClient.Create
I0917 16:56:29.890838 20150 start.go:167] duration metric: took 11.940274061s to libmachine.API.Create "addons-163060"
I0917 16:56:29.890853 20150 start.go:293] postStartSetup for "addons-163060" (driver="docker")
I0917 16:56:29.890866 20150 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0917 16:56:29.890930 20150 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0917 16:56:29.890990 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:29.907544 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:29.999136 20150 ssh_runner.go:195] Run: cat /etc/os-release
I0917 16:56:30.002070 20150 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0917 16:56:30.002095 20150 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0917 16:56:30.002111 20150 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0917 16:56:30.002118 20150 info.go:137] Remote host: Ubuntu 22.04.5 LTS
I0917 16:56:30.002130 20150 filesync.go:126] Scanning /home/jenkins/minikube-integration/19662-12004/.minikube/addons for local assets ...
I0917 16:56:30.002187 20150 filesync.go:126] Scanning /home/jenkins/minikube-integration/19662-12004/.minikube/files for local assets ...
I0917 16:56:30.002211 20150 start.go:296] duration metric: took 111.352738ms for postStartSetup
I0917 16:56:30.002478 20150 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-163060
I0917 16:56:30.018671 20150 profile.go:143] Saving config to /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/config.json ...
I0917 16:56:30.018906 20150 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0917 16:56:30.018944 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:30.034855 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:30.123548 20150 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0917 16:56:30.127394 20150 start.go:128] duration metric: took 12.180131932s to createHost
I0917 16:56:30.127420 20150 start.go:83] releasing machines lock for "addons-163060", held for 12.180267807s
I0917 16:56:30.127483 20150 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-163060
I0917 16:56:30.143030 20150 ssh_runner.go:195] Run: cat /version.json
I0917 16:56:30.143046 20150 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0917 16:56:30.143072 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:30.143099 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:30.159564 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:30.159577 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:30.246233 20150 ssh_runner.go:195] Run: systemctl --version
I0917 16:56:30.318995 20150 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0917 16:56:30.323074 20150 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0917 16:56:30.344528 20150 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0917 16:56:30.344604 20150 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0917 16:56:30.368056 20150 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0917 16:56:30.368079 20150 start.go:495] detecting cgroup driver to use...
I0917 16:56:30.368105 20150 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0917 16:56:30.368193 20150 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0917 16:56:30.381510 20150 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0917 16:56:30.389611 20150 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0917 16:56:30.397746 20150 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0917 16:56:30.397805 20150 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0917 16:56:30.406011 20150 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0917 16:56:30.414191 20150 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0917 16:56:30.422091 20150 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0917 16:56:30.430305 20150 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0917 16:56:30.437856 20150 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0917 16:56:30.445746 20150 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0917 16:56:30.453400 20150 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0917 16:56:30.461364 20150 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0917 16:56:30.467904 20150 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0917 16:56:30.474358 20150 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 16:56:30.547009 20150 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0917 16:56:30.620508 20150 start.go:495] detecting cgroup driver to use...
I0917 16:56:30.620554 20150 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0917 16:56:30.620611 20150 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0917 16:56:30.631834 20150 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0917 16:56:30.631894 20150 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0917 16:56:30.642568 20150 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0917 16:56:30.658145 20150 ssh_runner.go:195] Run: which cri-dockerd
I0917 16:56:30.661310 20150 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0917 16:56:30.669963 20150 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0917 16:56:30.686273 20150 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0917 16:56:30.779526 20150 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0917 16:56:30.869492 20150 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0917 16:56:30.869632 20150 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0917 16:56:30.885620 20150 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 16:56:30.963831 20150 ssh_runner.go:195] Run: sudo systemctl restart docker
I0917 16:56:31.209579 20150 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0917 16:56:31.219971 20150 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0917 16:56:31.230249 20150 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0917 16:56:31.307371 20150 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0917 16:56:31.372523 20150 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 16:56:31.444647 20150 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0917 16:56:31.456498 20150 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0917 16:56:31.466266 20150 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 16:56:31.536469 20150 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0917 16:56:31.595660 20150 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0917 16:56:31.595736 20150 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0917 16:56:31.598912 20150 start.go:563] Will wait 60s for crictl version
I0917 16:56:31.598964 20150 ssh_runner.go:195] Run: which crictl
I0917 16:56:31.602022 20150 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0917 16:56:31.632611 20150 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.2.1
RuntimeApiVersion: v1
I0917 16:56:31.632669 20150 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0917 16:56:31.655252 20150 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0917 16:56:31.679568 20150 out.go:235] * Preparing Kubernetes v1.31.1 on Docker 27.2.1 ...
I0917 16:56:31.679647 20150 cli_runner.go:164] Run: docker network inspect addons-163060 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0917 16:56:31.695986 20150 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0917 16:56:31.699375 20150 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0917 16:56:31.709335 20150 kubeadm.go:883] updating cluster {Name:addons-163060 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-163060 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0917 16:56:31.709487 20150 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0917 16:56:31.709546 20150 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0917 16:56:31.727647 20150 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0917 16:56:31.727669 20150 docker.go:615] Images already preloaded, skipping extraction
I0917 16:56:31.727713 20150 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0917 16:56:31.745244 20150 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0917 16:56:31.745271 20150 cache_images.go:84] Images are preloaded, skipping loading
I0917 16:56:31.745282 20150 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.1 docker true true} ...
I0917 16:56:31.745388 20150 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-163060 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.1 ClusterName:addons-163060 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0917 16:56:31.745455 20150 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0917 16:56:31.787082 20150 cni.go:84] Creating CNI manager for ""
I0917 16:56:31.787108 20150 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0917 16:56:31.787120 20150 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0917 16:56:31.787140 20150 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-163060 NodeName:addons-163060 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0917 16:56:31.787281 20150 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-163060"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0917 16:56:31.787336 20150 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.1
I0917 16:56:31.795157 20150 binaries.go:44] Found k8s binaries, skipping transfer
I0917 16:56:31.795217 20150 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0917 16:56:31.802487 20150 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0917 16:56:31.817343 20150 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0917 16:56:31.832712 20150 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0917 16:56:31.848161 20150 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0917 16:56:31.851254 20150 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0917 16:56:31.860684 20150 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 16:56:31.942358 20150 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0917 16:56:31.954317 20150 certs.go:68] Setting up /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060 for IP: 192.168.49.2
I0917 16:56:31.954338 20150 certs.go:194] generating shared ca certs ...
I0917 16:56:31.954357 20150 certs.go:226] acquiring lock for ca certs: {Name:mk4ca4c6226173ad89ccc5d68ab139f394e65c2b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:31.954477 20150 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19662-12004/.minikube/ca.key
I0917 16:56:32.210257 20150 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19662-12004/.minikube/ca.crt ...
I0917 16:56:32.210282 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/ca.crt: {Name:mkcd85584212dd22809c5b18f4d8bf6f30c0f290 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:32.210447 20150 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19662-12004/.minikube/ca.key ...
I0917 16:56:32.210457 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/ca.key: {Name:mkac9f869d50ca3b2cdbb23a9bdcaf276175f328 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:32.210525 20150 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19662-12004/.minikube/proxy-client-ca.key
I0917 16:56:32.614297 20150 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19662-12004/.minikube/proxy-client-ca.crt ...
I0917 16:56:32.614328 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/proxy-client-ca.crt: {Name:mked56b68cea3b3d464890eada72309bc1972dd4 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:32.614521 20150 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19662-12004/.minikube/proxy-client-ca.key ...
I0917 16:56:32.614535 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/proxy-client-ca.key: {Name:mk5bbfb3bcb7fa41ec180bd0390ac3aafc116b78 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:32.614627 20150 certs.go:256] generating profile certs ...
I0917 16:56:32.614696 20150 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/client.key
I0917 16:56:32.614725 20150 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/client.crt with IP's: []
I0917 16:56:32.916445 20150 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/client.crt ...
I0917 16:56:32.916474 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/client.crt: {Name:mk715bf18722721c21feb7cba118db233f5b44c5 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:32.916643 20150 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/client.key ...
I0917 16:56:32.916655 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/client.key: {Name:mk57b6fada5a9413c2083d5d507a5d4114b17dae Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:32.916726 20150 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.key.cbaf8388
I0917 16:56:32.916744 20150 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.crt.cbaf8388 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0917 16:56:33.026966 20150 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.crt.cbaf8388 ...
I0917 16:56:33.027011 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.crt.cbaf8388: {Name:mk7cbe5d81af6af0ed46a0d2d5badc08e21cddb8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:33.027178 20150 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.key.cbaf8388 ...
I0917 16:56:33.027191 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.key.cbaf8388: {Name:mk558d2263b705bda7be53e33b015fab9d3bf695 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:33.027260 20150 certs.go:381] copying /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.crt.cbaf8388 -> /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.crt
I0917 16:56:33.027333 20150 certs.go:385] copying /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.key.cbaf8388 -> /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.key
I0917 16:56:33.027375 20150 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/proxy-client.key
I0917 16:56:33.027393 20150 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/proxy-client.crt with IP's: []
I0917 16:56:33.189625 20150 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/proxy-client.crt ...
I0917 16:56:33.189652 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/proxy-client.crt: {Name:mk20a5e30ba07c3e5cd538b38350eef35358ea2c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:33.189801 20150 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/proxy-client.key ...
I0917 16:56:33.189811 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/proxy-client.key: {Name:mk951d9efe5308aefe8366c2fa40fbbfc872c9d3 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:33.189970 20150 certs.go:484] found cert: /home/jenkins/minikube-integration/19662-12004/.minikube/certs/ca-key.pem (1675 bytes)
I0917 16:56:33.190004 20150 certs.go:484] found cert: /home/jenkins/minikube-integration/19662-12004/.minikube/certs/ca.pem (1082 bytes)
I0917 16:56:33.190027 20150 certs.go:484] found cert: /home/jenkins/minikube-integration/19662-12004/.minikube/certs/cert.pem (1123 bytes)
I0917 16:56:33.190048 20150 certs.go:484] found cert: /home/jenkins/minikube-integration/19662-12004/.minikube/certs/key.pem (1679 bytes)
I0917 16:56:33.190647 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0917 16:56:33.212184 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0917 16:56:33.233677 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0917 16:56:33.254791 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0917 16:56:33.275287 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0917 16:56:33.295588 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0917 16:56:33.315968 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0917 16:56:33.336105 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/profiles/addons-163060/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1679 bytes)
I0917 16:56:33.355999 20150 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19662-12004/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0917 16:56:33.376341 20150 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0917 16:56:33.391105 20150 ssh_runner.go:195] Run: openssl version
I0917 16:56:33.395774 20150 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0917 16:56:33.403910 20150 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0917 16:56:33.406963 20150 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 17 16:56 /usr/share/ca-certificates/minikubeCA.pem
I0917 16:56:33.407027 20150 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0917 16:56:33.413750 20150 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0917 16:56:33.422203 20150 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0917 16:56:33.425364 20150 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0917 16:56:33.425410 20150 kubeadm.go:392] StartCluster: {Name:addons-163060 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726589491-19662@sha256:6370b9fec173944088c2d87d44b01819c0ec611a83d9e2f38d36352dff8121a4 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-163060 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0917 16:56:33.425534 20150 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0917 16:56:33.443333 20150 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0917 16:56:33.452025 20150 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0917 16:56:33.459386 20150 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0917 16:56:33.459445 20150 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0917 16:56:33.466478 20150 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0917 16:56:33.466492 20150 kubeadm.go:157] found existing configuration files:
I0917 16:56:33.466522 20150 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0917 16:56:33.473909 20150 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0917 16:56:33.473960 20150 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0917 16:56:33.480958 20150 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0917 16:56:33.487882 20150 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0917 16:56:33.487920 20150 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0917 16:56:33.494930 20150 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0917 16:56:33.502056 20150 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0917 16:56:33.502098 20150 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0917 16:56:33.508925 20150 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0917 16:56:33.516081 20150 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0917 16:56:33.516119 20150 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0917 16:56:33.522923 20150 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0917 16:56:33.556343 20150 kubeadm.go:310] [init] Using Kubernetes version: v1.31.1
I0917 16:56:33.556429 20150 kubeadm.go:310] [preflight] Running pre-flight checks
I0917 16:56:33.574365 20150 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0917 16:56:33.574442 20150 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1069-gcp[0m
I0917 16:56:33.574511 20150 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0917 16:56:33.574602 20150 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0917 16:56:33.574677 20150 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0917 16:56:33.574749 20150 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0917 16:56:33.574826 20150 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0917 16:56:33.574875 20150 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0917 16:56:33.574918 20150 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0917 16:56:33.574961 20150 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0917 16:56:33.575030 20150 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0917 16:56:33.575075 20150 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0917 16:56:33.622617 20150 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0917 16:56:33.622757 20150 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0917 16:56:33.622901 20150 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0917 16:56:33.632420 20150 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0917 16:56:33.635235 20150 out.go:235] - Generating certificates and keys ...
I0917 16:56:33.635344 20150 kubeadm.go:310] [certs] Using existing ca certificate authority
I0917 16:56:33.635412 20150 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0917 16:56:33.897665 20150 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0917 16:56:34.231585 20150 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0917 16:56:34.506446 20150 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0917 16:56:34.664847 20150 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0917 16:56:34.753535 20150 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0917 16:56:34.753659 20150 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-163060 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0917 16:56:35.125631 20150 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0917 16:56:35.125768 20150 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-163060 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0917 16:56:35.414806 20150 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0917 16:56:35.551977 20150 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0917 16:56:35.720509 20150 kubeadm.go:310] [certs] Generating "sa" key and public key
I0917 16:56:35.720586 20150 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0917 16:56:35.961015 20150 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0917 16:56:36.164219 20150 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0917 16:56:36.317872 20150 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0917 16:56:36.552734 20150 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0917 16:56:36.652410 20150 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0917 16:56:36.652771 20150 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0917 16:56:36.655149 20150 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0917 16:56:36.657223 20150 out.go:235] - Booting up control plane ...
I0917 16:56:36.657359 20150 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0917 16:56:36.657448 20150 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0917 16:56:36.657505 20150 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0917 16:56:36.665959 20150 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0917 16:56:36.671265 20150 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0917 16:56:36.671330 20150 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0917 16:56:36.753740 20150 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0917 16:56:36.753923 20150 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0917 16:56:37.755069 20150 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 1.001343664s
I0917 16:56:37.755200 20150 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0917 16:56:41.756540 20150 kubeadm.go:310] [api-check] The API server is healthy after 4.001525789s
I0917 16:56:41.768130 20150 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0917 16:56:41.779673 20150 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0917 16:56:41.794250 20150 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0917 16:56:41.794526 20150 kubeadm.go:310] [mark-control-plane] Marking the node addons-163060 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0917 16:56:41.801108 20150 kubeadm.go:310] [bootstrap-token] Using token: hbcvk0.770ivrsaqb5vjd99
I0917 16:56:41.802554 20150 out.go:235] - Configuring RBAC rules ...
I0917 16:56:41.802702 20150 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0917 16:56:41.805524 20150 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0917 16:56:41.811609 20150 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0917 16:56:41.814193 20150 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0917 16:56:41.816587 20150 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0917 16:56:41.818630 20150 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0917 16:56:42.162174 20150 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0917 16:56:42.588059 20150 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0917 16:56:43.161784 20150 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0917 16:56:43.163099 20150 kubeadm.go:310]
I0917 16:56:43.163258 20150 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0917 16:56:43.163270 20150 kubeadm.go:310]
I0917 16:56:43.163368 20150 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0917 16:56:43.163377 20150 kubeadm.go:310]
I0917 16:56:43.163413 20150 kubeadm.go:310] mkdir -p $HOME/.kube
I0917 16:56:43.163507 20150 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0917 16:56:43.163584 20150 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0917 16:56:43.163592 20150 kubeadm.go:310]
I0917 16:56:43.163667 20150 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0917 16:56:43.163676 20150 kubeadm.go:310]
I0917 16:56:43.163754 20150 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0917 16:56:43.163763 20150 kubeadm.go:310]
I0917 16:56:43.163845 20150 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0917 16:56:43.163953 20150 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0917 16:56:43.164051 20150 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0917 16:56:43.164059 20150 kubeadm.go:310]
I0917 16:56:43.164186 20150 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0917 16:56:43.164300 20150 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0917 16:56:43.164311 20150 kubeadm.go:310]
I0917 16:56:43.164414 20150 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token hbcvk0.770ivrsaqb5vjd99 \
I0917 16:56:43.164564 20150 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:39ed9eaf7eb785cbd3eb1cdde44a12cb8b03a8530d08ba6cf08757c75b478eb2 \
I0917 16:56:43.164598 20150 kubeadm.go:310] --control-plane
I0917 16:56:43.164607 20150 kubeadm.go:310]
I0917 16:56:43.164714 20150 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0917 16:56:43.164722 20150 kubeadm.go:310]
I0917 16:56:43.164830 20150 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token hbcvk0.770ivrsaqb5vjd99 \
I0917 16:56:43.164965 20150 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:39ed9eaf7eb785cbd3eb1cdde44a12cb8b03a8530d08ba6cf08757c75b478eb2
I0917 16:56:43.166763 20150 kubeadm.go:310] W0917 16:56:33.553909 1929 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0917 16:56:43.167195 20150 kubeadm.go:310] W0917 16:56:33.554517 1929 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0917 16:56:43.167545 20150 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1069-gcp\n", err: exit status 1
I0917 16:56:43.167728 20150 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0917 16:56:43.167746 20150 cni.go:84] Creating CNI manager for ""
I0917 16:56:43.167763 20150 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0917 16:56:43.169703 20150 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0917 16:56:43.171054 20150 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0917 16:56:43.179013 20150 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0917 16:56:43.194659 20150 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0917 16:56:43.194734 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:43.194754 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-163060 minikube.k8s.io/updated_at=2024_09_17T16_56_43_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=825de77780746e57a7948604e1eea9da920a46ce minikube.k8s.io/name=addons-163060 minikube.k8s.io/primary=true
I0917 16:56:43.201390 20150 ops.go:34] apiserver oom_adj: -16
I0917 16:56:43.262928 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:43.763127 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:44.263092 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:44.763569 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:45.263637 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:45.763995 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:46.263767 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:46.763326 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:47.263598 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:47.763139 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:48.263102 20150 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0917 16:56:48.358854 20150 kubeadm.go:1113] duration metric: took 5.164174214s to wait for elevateKubeSystemPrivileges
I0917 16:56:48.358890 20150 kubeadm.go:394] duration metric: took 14.93348494s to StartCluster
I0917 16:56:48.358908 20150 settings.go:142] acquiring lock: {Name:mkb8576b2f39f9923d5cc12f8cc85696a352bae4 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:48.359023 20150 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19662-12004/kubeconfig
I0917 16:56:48.359360 20150 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19662-12004/kubeconfig: {Name:mk0b336c0df7435007d298d52b4ddaa46513b06b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0917 16:56:48.359522 20150 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0917 16:56:48.359537 20150 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0917 16:56:48.359595 20150 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0917 16:56:48.359717 20150 addons.go:69] Setting yakd=true in profile "addons-163060"
I0917 16:56:48.359737 20150 addons.go:234] Setting addon yakd=true in "addons-163060"
I0917 16:56:48.359740 20150 addons.go:69] Setting gcp-auth=true in profile "addons-163060"
I0917 16:56:48.359743 20150 config.go:182] Loaded profile config "addons-163060": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0917 16:56:48.359755 20150 addons.go:69] Setting ingress-dns=true in profile "addons-163060"
I0917 16:56:48.359766 20150 mustload.go:65] Loading cluster: addons-163060
I0917 16:56:48.359771 20150 addons.go:69] Setting helm-tiller=true in profile "addons-163060"
I0917 16:56:48.359773 20150 addons.go:69] Setting storage-provisioner=true in profile "addons-163060"
I0917 16:56:48.359785 20150 addons.go:234] Setting addon ingress-dns=true in "addons-163060"
I0917 16:56:48.359789 20150 addons.go:69] Setting volcano=true in profile "addons-163060"
I0917 16:56:48.359792 20150 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-163060"
I0917 16:56:48.359800 20150 addons.go:69] Setting metrics-server=true in profile "addons-163060"
I0917 16:56:48.359806 20150 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-163060"
I0917 16:56:48.359810 20150 addons.go:69] Setting volumesnapshots=true in profile "addons-163060"
I0917 16:56:48.359812 20150 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-163060"
I0917 16:56:48.359815 20150 addons.go:234] Setting addon metrics-server=true in "addons-163060"
I0917 16:56:48.359820 20150 addons.go:234] Setting addon volumesnapshots=true in "addons-163060"
I0917 16:56:48.359826 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.359825 20150 addons.go:69] Setting default-storageclass=true in profile "addons-163060"
I0917 16:56:48.359836 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.359841 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.359840 20150 addons.go:69] Setting cloud-spanner=true in profile "addons-163060"
I0917 16:56:48.359846 20150 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-163060"
I0917 16:56:48.359781 20150 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-163060"
I0917 16:56:48.359855 20150 addons.go:234] Setting addon cloud-spanner=true in "addons-163060"
I0917 16:56:48.359860 20150 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-163060"
I0917 16:56:48.359871 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.359882 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.359942 20150 config.go:182] Loaded profile config "addons-163060": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0917 16:56:48.359773 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.359754 20150 addons.go:69] Setting ingress=true in profile "addons-163060"
I0917 16:56:48.359985 20150 addons.go:234] Setting addon ingress=true in "addons-163060"
I0917 16:56:48.360007 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.360216 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.360257 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.360357 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.360370 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.360386 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.360410 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.360426 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.360520 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.359733 20150 addons.go:69] Setting inspektor-gadget=true in profile "addons-163060"
I0917 16:56:48.360606 20150 addons.go:234] Setting addon inspektor-gadget=true in "addons-163060"
I0917 16:56:48.360675 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.359790 20150 addons.go:234] Setting addon helm-tiller=true in "addons-163060"
I0917 16:56:48.360918 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.361130 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.361305 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.359802 20150 addons.go:234] Setting addon volcano=true in "addons-163060"
I0917 16:56:48.361345 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.359830 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.359845 20150 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-163060"
I0917 16:56:48.361422 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.359802 20150 addons.go:69] Setting registry=true in profile "addons-163060"
I0917 16:56:48.361817 20150 addons.go:234] Setting addon registry=true in "addons-163060"
I0917 16:56:48.361848 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.359794 20150 addons.go:234] Setting addon storage-provisioner=true in "addons-163060"
I0917 16:56:48.362156 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.363818 20150 out.go:177] * Verifying Kubernetes components...
I0917 16:56:48.365228 20150 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0917 16:56:48.379628 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.379628 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.380081 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.380470 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.380951 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.393958 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.408574 20150 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-163060"
I0917 16:56:48.408624 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.409176 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.414622 20150 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0917 16:56:48.414757 20150 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0917 16:56:48.416323 20150 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0917 16:56:48.416343 20150 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0917 16:56:48.416412 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.416694 20150 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0917 16:56:48.416707 20150 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0917 16:56:48.416769 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.443127 20150 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0917 16:56:48.443127 20150 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0917 16:56:48.447260 20150 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0917 16:56:48.447280 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0917 16:56:48.447336 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.455044 20150 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0917 16:56:48.457009 20150 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0917 16:56:48.458444 20150 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0917 16:56:48.458769 20150 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0917 16:56:48.458786 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0917 16:56:48.458852 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.461013 20150 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0917 16:56:48.461124 20150 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0917 16:56:48.461169 20150 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0917 16:56:48.462311 20150 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0917 16:56:48.462324 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0917 16:56:48.462374 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.462829 20150 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0917 16:56:48.464049 20150 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0917 16:56:48.464132 20150 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0917 16:56:48.465248 20150 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0917 16:56:48.466045 20150 addons.go:234] Setting addon default-storageclass=true in "addons-163060"
I0917 16:56:48.466081 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:48.466869 20150 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0917 16:56:48.466885 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0917 16:56:48.467071 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.467373 20150 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0917 16:56:48.467386 20150 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0917 16:56:48.467429 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.467731 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:48.469342 20150 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0917 16:56:48.470124 20150 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0917 16:56:48.472364 20150 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0917 16:56:48.472475 20150 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0917 16:56:48.474568 20150 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0917 16:56:48.474676 20150 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0917 16:56:48.474893 20150 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0917 16:56:48.477537 20150 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0917 16:56:48.477556 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0917 16:56:48.477606 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.478556 20150 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0917 16:56:48.478576 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0917 16:56:48.478626 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.478791 20150 out.go:177] - Using image docker.io/registry:2.8.3
I0917 16:56:48.478939 20150 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0917 16:56:48.478951 20150 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0917 16:56:48.479029 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.481017 20150 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0917 16:56:48.482648 20150 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0917 16:56:48.482664 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0917 16:56:48.482717 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.484518 20150 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0917 16:56:48.485861 20150 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0917 16:56:48.485980 20150 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0917 16:56:48.487031 20150 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0917 16:56:48.487048 20150 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0917 16:56:48.487104 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.487232 20150 out.go:177] - Using image docker.io/busybox:stable
I0917 16:56:48.487362 20150 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0917 16:56:48.487372 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0917 16:56:48.487411 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.487437 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.488485 20150 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0917 16:56:48.488498 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0917 16:56:48.488535 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.508389 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.517938 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.518882 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.531973 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.536535 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.536943 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.539073 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.539266 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.540561 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.541334 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.541896 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.542667 20150 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0917 16:56:48.542685 20150 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0917 16:56:48.542729 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:48.544346 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.546554 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
W0917 16:56:48.557913 20150 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0917 16:56:48.557944 20150 retry.go:31] will retry after 305.266352ms: ssh: handshake failed: EOF
W0917 16:56:48.563427 20150 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0917 16:56:48.563460 20150 retry.go:31] will retry after 231.410259ms: ssh: handshake failed: EOF
I0917 16:56:48.570171 20150 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0917 16:56:48.570385 20150 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0917 16:56:48.579328 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:48.959141 20150 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0917 16:56:48.959184 20150 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0917 16:56:48.963403 20150 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0917 16:56:48.963428 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0917 16:56:48.970601 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0917 16:56:49.054996 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0917 16:56:49.064512 20150 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0917 16:56:49.064558 20150 ssh_runner.go:362] scp helm-tiller/helm-tiller-rbac.yaml --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0917 16:56:49.064783 20150 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0917 16:56:49.064836 20150 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0917 16:56:49.071741 20150 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0917 16:56:49.071764 20150 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0917 16:56:49.162242 20150 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0917 16:56:49.162273 20150 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0917 16:56:49.168551 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0917 16:56:49.255819 20150 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0917 16:56:49.255846 20150 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0917 16:56:49.256258 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0917 16:56:49.261469 20150 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0917 16:56:49.261545 20150 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0917 16:56:49.262759 20150 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0917 16:56:49.262808 20150 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0917 16:56:49.347632 20150 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0917 16:56:49.347676 20150 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0917 16:56:49.351393 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0917 16:56:49.353648 20150 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0917 16:56:49.353666 20150 ssh_runner.go:362] scp helm-tiller/helm-tiller-svc.yaml --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0917 16:56:49.450098 20150 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0917 16:56:49.450202 20150 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0917 16:56:49.453620 20150 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0917 16:56:49.453700 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0917 16:56:49.455735 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0917 16:56:49.551832 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0917 16:56:49.556307 20150 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0917 16:56:49.556339 20150 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0917 16:56:49.648680 20150 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0917 16:56:49.648713 20150 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0917 16:56:49.654532 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0917 16:56:49.756357 20150 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0917 16:56:49.756442 20150 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0917 16:56:49.759199 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0917 16:56:49.848389 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0917 16:56:49.959405 20150 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.38899137s)
I0917 16:56:49.959496 20150 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0917 16:56:49.959674 20150 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.389476446s)
I0917 16:56:49.961771 20150 node_ready.go:35] waiting up to 6m0s for node "addons-163060" to be "Ready" ...
I0917 16:56:49.962388 20150 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0917 16:56:49.962450 20150 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0917 16:56:49.962605 20150 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0917 16:56:49.962640 20150 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0917 16:56:49.965946 20150 node_ready.go:49] node "addons-163060" has status "Ready":"True"
I0917 16:56:49.965969 20150 node_ready.go:38] duration metric: took 4.054358ms for node "addons-163060" to be "Ready" ...
I0917 16:56:49.965981 20150 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0917 16:56:49.976858 20150 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-f8spg" in "kube-system" namespace to be "Ready" ...
I0917 16:56:50.056415 20150 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0917 16:56:50.056443 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0917 16:56:50.255693 20150 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0917 16:56:50.255718 20150 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0917 16:56:50.453239 20150 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0917 16:56:50.453323 20150 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0917 16:56:50.464471 20150 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-163060" context rescaled to 1 replicas
I0917 16:56:50.549453 20150 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0917 16:56:50.549531 20150 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0917 16:56:50.551041 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (1.580401766s)
I0917 16:56:50.654594 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0917 16:56:50.669594 20150 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0917 16:56:50.669621 20150 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0917 16:56:50.750105 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0917 16:56:51.054042 20150 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0917 16:56:51.054067 20150 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0917 16:56:51.056758 20150 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0917 16:56:51.056783 20150 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0917 16:56:51.657322 20150 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0917 16:56:51.657398 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0917 16:56:51.672350 20150 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0917 16:56:51.672429 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0917 16:56:51.850850 20150 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0917 16:56:51.851015 20150 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0917 16:56:52.052032 20150 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0917 16:56:52.052120 20150 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0917 16:56:52.062254 20150 pod_ready.go:103] pod "coredns-7c65d6cfc9-f8spg" in "kube-system" namespace has status "Ready":"False"
I0917 16:56:52.251758 20150 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0917 16:56:52.251789 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0917 16:56:52.259783 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (3.204752477s)
I0917 16:56:52.259753 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (3.091096436s)
I0917 16:56:52.368210 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0917 16:56:52.449313 20150 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0917 16:56:52.449361 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0917 16:56:52.948868 20150 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0917 16:56:52.948909 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0917 16:56:52.951058 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0917 16:56:53.351553 20150 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0917 16:56:53.351585 20150 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0917 16:56:54.149974 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0917 16:56:54.553681 20150 pod_ready.go:103] pod "coredns-7c65d6cfc9-f8spg" in "kube-system" namespace has status "Ready":"False"
I0917 16:56:55.062302 20150 pod_ready.go:93] pod "coredns-7c65d6cfc9-f8spg" in "kube-system" namespace has status "Ready":"True"
I0917 16:56:55.062337 20150 pod_ready.go:82] duration metric: took 5.085453472s for pod "coredns-7c65d6cfc9-f8spg" in "kube-system" namespace to be "Ready" ...
I0917 16:56:55.062350 20150 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-k87c9" in "kube-system" namespace to be "Ready" ...
I0917 16:56:55.453246 20150 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0917 16:56:55.453337 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:55.485414 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:56.257437 20150 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0917 16:56:56.469795 20150 addons.go:234] Setting addon gcp-auth=true in "addons-163060"
I0917 16:56:56.469860 20150 host.go:66] Checking if "addons-163060" exists ...
I0917 16:56:56.470377 20150 cli_runner.go:164] Run: docker container inspect addons-163060 --format={{.State.Status}}
I0917 16:56:56.487308 20150 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0917 16:56:56.487363 20150 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-163060
I0917 16:56:56.503000 20150 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19662-12004/.minikube/machines/addons-163060/id_rsa Username:docker}
I0917 16:56:57.069194 20150 pod_ready.go:103] pod "coredns-7c65d6cfc9-k87c9" in "kube-system" namespace has status "Ready":"False"
I0917 16:56:59.072816 20150 pod_ready.go:103] pod "coredns-7c65d6cfc9-k87c9" in "kube-system" namespace has status "Ready":"False"
I0917 16:57:00.451389 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (11.0999625s)
I0917 16:57:00.451311 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.195013128s)
I0917 16:57:00.451503 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (10.995685384s)
I0917 16:57:00.451870 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (10.900008105s)
I0917 16:57:00.451923 20150 addons.go:475] Verifying addon ingress=true in "addons-163060"
I0917 16:57:00.452212 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (10.797598726s)
I0917 16:57:00.452303 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (10.693042573s)
I0917 16:57:00.452354 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (10.603883056s)
I0917 16:57:00.452907 20150 addons.go:475] Verifying addon registry=true in "addons-163060"
I0917 16:57:00.452401 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (9.797721945s)
I0917 16:57:00.452472 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (9.702336214s)
I0917 16:57:00.453154 20150 addons.go:475] Verifying addon metrics-server=true in "addons-163060"
I0917 16:57:00.452576 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (8.084333174s)
W0917 16:57:00.453199 20150 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0917 16:57:00.453214 20150 retry.go:31] will retry after 309.290798ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0917 16:57:00.452651 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (7.501565854s)
I0917 16:57:00.455019 20150 out.go:177] * Verifying ingress addon...
I0917 16:57:00.455022 20150 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-163060 service yakd-dashboard -n yakd-dashboard
I0917 16:57:00.456141 20150 out.go:177] * Verifying registry addon...
I0917 16:57:00.461253 20150 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0917 16:57:00.461256 20150 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0917 16:57:00.467306 20150 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0917 16:57:00.467332 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:00.467760 20150 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0917 16:57:00.467783 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:00.663779 20150 pod_ready.go:98] pod "coredns-7c65d6cfc9-k87c9" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-17 16:57:00 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-17 16:56:48 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-17 16:56:48 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-17 16:56:48 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-17 16:56:48 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.49.2 HostIPs:[{IP:192.168.49.2
}] PodIP:10.244.0.2 PodIPs:[{IP:10.244.0.2}] StartTime:2024-09-17 16:56:48 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-09-17 16:56:50 +0000 UTC,FinishedAt:2024-09-17 16:56:58 +0000 UTC,ContainerID:docker://6adc028f05fabe97df4091159f08bb34f70f3dceacb1c09e4a49b77933cd5f3e,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.3 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e ContainerID:docker://6adc028f05fabe97df4091159f08bb34f70f3dceacb1c09e4a49b77933cd5f3e Started:0xc00208a500 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0xc001abff60} {Name:kube-api-access-ff2pg MountPath:/var/run/secrets/kubernetes.io/serviceaccount
ReadOnly:true RecursiveReadOnly:0xc001abff70}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
I0917 16:57:00.663909 20150 pod_ready.go:82] duration metric: took 5.601549044s for pod "coredns-7c65d6cfc9-k87c9" in "kube-system" namespace to be "Ready" ...
E0917 16:57:00.663939 20150 pod_ready.go:67] WaitExtra: waitPodCondition: pod "coredns-7c65d6cfc9-k87c9" in "kube-system" namespace has status phase "Succeeded" (skipping!): {Phase:Succeeded Conditions:[{Type:PodReadyToStartContainers Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-17 16:57:00 +0000 UTC Reason: Message:} {Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-17 16:56:48 +0000 UTC Reason:PodCompleted Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-17 16:56:48 +0000 UTC Reason:PodCompleted Message:} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-17 16:56:48 +0000 UTC Reason:PodCompleted Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2024-09-17 16:56:48 +0000 UTC Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:192.168.4
9.2 HostIPs:[{IP:192.168.49.2}] PodIP:10.244.0.2 PodIPs:[{IP:10.244.0.2}] StartTime:2024-09-17 16:56:48 +0000 UTC InitContainerStatuses:[] ContainerStatuses:[{Name:coredns State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2024-09-17 16:56:50 +0000 UTC,FinishedAt:2024-09-17 16:56:58 +0000 UTC,ContainerID:docker://6adc028f05fabe97df4091159f08bb34f70f3dceacb1c09e4a49b77933cd5f3e,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:registry.k8s.io/coredns/coredns:v1.11.3 ImageID:docker-pullable://registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e ContainerID:docker://6adc028f05fabe97df4091159f08bb34f70f3dceacb1c09e4a49b77933cd5f3e Started:0xc00208a500 AllocatedResources:map[] Resources:nil VolumeMounts:[{Name:config-volume MountPath:/etc/coredns ReadOnly:true RecursiveReadOnly:0xc001abff60} {Name:kube-api-access-ff2pg MountPath:/var/run/secrets
/kubernetes.io/serviceaccount ReadOnly:true RecursiveReadOnly:0xc001abff70}] User:nil AllocatedResourcesStatus:[]}] QOSClass:Burstable EphemeralContainerStatuses:[] Resize: ResourceClaimStatuses:[]}
I0917 16:57:00.663960 20150 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-163060" in "kube-system" namespace to be "Ready" ...
I0917 16:57:00.756779 20150 pod_ready.go:93] pod "etcd-addons-163060" in "kube-system" namespace has status "Ready":"True"
I0917 16:57:00.756867 20150 pod_ready.go:82] duration metric: took 92.885407ms for pod "etcd-addons-163060" in "kube-system" namespace to be "Ready" ...
I0917 16:57:00.756897 20150 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-163060" in "kube-system" namespace to be "Ready" ...
I0917 16:57:00.763598 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0917 16:57:00.852480 20150 pod_ready.go:93] pod "kube-apiserver-addons-163060" in "kube-system" namespace has status "Ready":"True"
I0917 16:57:00.852571 20150 pod_ready.go:82] duration metric: took 95.653344ms for pod "kube-apiserver-addons-163060" in "kube-system" namespace to be "Ready" ...
I0917 16:57:00.852600 20150 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-163060" in "kube-system" namespace to be "Ready" ...
I0917 16:57:00.859572 20150 pod_ready.go:93] pod "kube-controller-manager-addons-163060" in "kube-system" namespace has status "Ready":"True"
I0917 16:57:00.859648 20150 pod_ready.go:82] duration metric: took 7.028464ms for pod "kube-controller-manager-addons-163060" in "kube-system" namespace to be "Ready" ...
I0917 16:57:00.859674 20150 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-9xj99" in "kube-system" namespace to be "Ready" ...
I0917 16:57:00.867948 20150 pod_ready.go:93] pod "kube-proxy-9xj99" in "kube-system" namespace has status "Ready":"True"
I0917 16:57:00.867980 20150 pod_ready.go:82] duration metric: took 8.287251ms for pod "kube-proxy-9xj99" in "kube-system" namespace to be "Ready" ...
I0917 16:57:00.867993 20150 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-163060" in "kube-system" namespace to be "Ready" ...
I0917 16:57:00.965323 20150 pod_ready.go:93] pod "kube-scheduler-addons-163060" in "kube-system" namespace has status "Ready":"True"
I0917 16:57:00.965365 20150 pod_ready.go:82] duration metric: took 97.361861ms for pod "kube-scheduler-addons-163060" in "kube-system" namespace to be "Ready" ...
I0917 16:57:00.965375 20150 pod_ready.go:39] duration metric: took 10.999334441s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0917 16:57:00.965401 20150 api_server.go:52] waiting for apiserver process to appear ...
I0917 16:57:00.965506 20150 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0917 16:57:00.965633 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:00.967443 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:01.468095 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:01.469325 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:01.654093 20150 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.166744622s)
I0917 16:57:01.654091 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (7.503844666s)
I0917 16:57:01.654324 20150 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-163060"
I0917 16:57:01.656939 20150 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0917 16:57:01.656954 20150 out.go:177] * Verifying csi-hostpath-driver addon...
I0917 16:57:01.658727 20150 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0917 16:57:01.659590 20150 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0917 16:57:01.660953 20150 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0917 16:57:01.661003 20150 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0917 16:57:01.666058 20150 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0917 16:57:01.666124 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:01.761915 20150 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0917 16:57:01.761998 20150 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0917 16:57:01.856831 20150 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0917 16:57:01.856858 20150 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0917 16:57:01.947822 20150 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0917 16:57:01.965625 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:01.967455 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:02.166119 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:02.466769 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:02.466831 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:02.667023 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:02.966195 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:02.966903 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:03.166367 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:03.451791 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.688080518s)
I0917 16:57:03.451819 20150 ssh_runner.go:235] Completed: sudo pgrep -xnf kube-apiserver.*minikube.*: (2.486276073s)
I0917 16:57:03.451903 20150 api_server.go:72] duration metric: took 15.092335753s to wait for apiserver process to appear ...
I0917 16:57:03.451908 20150 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.504007862s)
I0917 16:57:03.451915 20150 api_server.go:88] waiting for apiserver healthz status ...
I0917 16:57:03.452162 20150 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0917 16:57:03.453466 20150 addons.go:475] Verifying addon gcp-auth=true in "addons-163060"
I0917 16:57:03.455394 20150 out.go:177] * Verifying gcp-auth addon...
I0917 16:57:03.457639 20150 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0917 16:57:03.458055 20150 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0917 16:57:03.458819 20150 api_server.go:141] control plane version: v1.31.1
I0917 16:57:03.458841 20150 api_server.go:131] duration metric: took 6.70063ms to wait for apiserver health ...
I0917 16:57:03.458851 20150 system_pods.go:43] waiting for kube-system pods to appear ...
I0917 16:57:03.459717 20150 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0917 16:57:03.464302 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:03.464536 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:03.467502 20150 system_pods.go:59] 18 kube-system pods found
I0917 16:57:03.467542 20150 system_pods.go:61] "coredns-7c65d6cfc9-f8spg" [e61f3b7d-af39-440c-a80e-dc94ddb90c07] Running
I0917 16:57:03.467560 20150 system_pods.go:61] "csi-hostpath-attacher-0" [b2f35197-efe2-4181-ad9d-4fc3a1c5cb47] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0917 16:57:03.467572 20150 system_pods.go:61] "csi-hostpath-resizer-0" [48a87fd7-c4de-4b57-b089-3d3cd4d802ea] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0917 16:57:03.467587 20150 system_pods.go:61] "csi-hostpathplugin-kfwr4" [4dad6b37-811d-40dd-81e3-f10b2e9d00b9] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0917 16:57:03.467597 20150 system_pods.go:61] "etcd-addons-163060" [ab42a568-f4f1-4a0d-964d-030dba5c9eef] Running
I0917 16:57:03.467604 20150 system_pods.go:61] "kube-apiserver-addons-163060" [e849fea6-23da-4f64-a864-d5e83d161f5f] Running
I0917 16:57:03.467612 20150 system_pods.go:61] "kube-controller-manager-addons-163060" [66594d00-9610-4b6f-861f-4e049879121e] Running
I0917 16:57:03.467620 20150 system_pods.go:61] "kube-ingress-dns-minikube" [315f956b-f0bb-433b-9cc2-55581bcebdd4] Pending / Ready:ContainersNotReady (containers with unready status: [minikube-ingress-dns]) / ContainersReady:ContainersNotReady (containers with unready status: [minikube-ingress-dns])
I0917 16:57:03.467628 20150 system_pods.go:61] "kube-proxy-9xj99" [94e38d82-f813-4c73-ad0f-2b1d5bfd1a97] Running
I0917 16:57:03.467636 20150 system_pods.go:61] "kube-scheduler-addons-163060" [1a1eaf4b-4adc-4a9f-bcf3-2f8a56738f62] Running
I0917 16:57:03.467646 20150 system_pods.go:61] "metrics-server-84c5f94fbc-2f2f2" [03a25efb-5c8d-4637-b228-6bb67ccb601f] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0917 16:57:03.467658 20150 system_pods.go:61] "nvidia-device-plugin-daemonset-fvg2d" [69980d79-6040-46a6-92e4-f154f528e261] Pending / Ready:ContainersNotReady (containers with unready status: [nvidia-device-plugin-ctr]) / ContainersReady:ContainersNotReady (containers with unready status: [nvidia-device-plugin-ctr])
I0917 16:57:03.467676 20150 system_pods.go:61] "registry-66c9cd494c-xnftt" [87171e43-6b56-423a-ac20-6b46a3583197] Pending / Ready:ContainersNotReady (containers with unready status: [registry]) / ContainersReady:ContainersNotReady (containers with unready status: [registry])
I0917 16:57:03.467689 20150 system_pods.go:61] "registry-proxy-9ztsk" [de43c7a6-1992-4444-969d-d41949e06cdb] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0917 16:57:03.467700 20150 system_pods.go:61] "snapshot-controller-56fcc65765-fzkt2" [db3a7532-a6e9-481f-9449-e5e5f81fb4db] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0917 16:57:03.467711 20150 system_pods.go:61] "snapshot-controller-56fcc65765-t7gvk" [6e3f0826-ee2e-4a50-9797-2c8c8954b6a8] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0917 16:57:03.467717 20150 system_pods.go:61] "storage-provisioner" [bbad410c-6bb7-404e-bfd5-cb7d0e8f806c] Running
I0917 16:57:03.467727 20150 system_pods.go:61] "tiller-deploy-b48cc5f79-qd92x" [6de636f4-5713-4439-9d76-756777a66ef2] Pending / Ready:ContainersNotReady (containers with unready status: [tiller]) / ContainersReady:ContainersNotReady (containers with unready status: [tiller])
I0917 16:57:03.467735 20150 system_pods.go:74] duration metric: took 8.877494ms to wait for pod list to return data ...
I0917 16:57:03.467747 20150 default_sa.go:34] waiting for default service account to be created ...
I0917 16:57:03.469710 20150 default_sa.go:45] found service account: "default"
I0917 16:57:03.469726 20150 default_sa.go:55] duration metric: took 1.973525ms for default service account to be created ...
I0917 16:57:03.469735 20150 system_pods.go:116] waiting for k8s-apps to be running ...
I0917 16:57:03.481005 20150 system_pods.go:86] 18 kube-system pods found
I0917 16:57:03.481031 20150 system_pods.go:89] "coredns-7c65d6cfc9-f8spg" [e61f3b7d-af39-440c-a80e-dc94ddb90c07] Running
I0917 16:57:03.481040 20150 system_pods.go:89] "csi-hostpath-attacher-0" [b2f35197-efe2-4181-ad9d-4fc3a1c5cb47] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0917 16:57:03.481046 20150 system_pods.go:89] "csi-hostpath-resizer-0" [48a87fd7-c4de-4b57-b089-3d3cd4d802ea] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0917 16:57:03.481057 20150 system_pods.go:89] "csi-hostpathplugin-kfwr4" [4dad6b37-811d-40dd-81e3-f10b2e9d00b9] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0917 16:57:03.481065 20150 system_pods.go:89] "etcd-addons-163060" [ab42a568-f4f1-4a0d-964d-030dba5c9eef] Running
I0917 16:57:03.481069 20150 system_pods.go:89] "kube-apiserver-addons-163060" [e849fea6-23da-4f64-a864-d5e83d161f5f] Running
I0917 16:57:03.481073 20150 system_pods.go:89] "kube-controller-manager-addons-163060" [66594d00-9610-4b6f-861f-4e049879121e] Running
I0917 16:57:03.481082 20150 system_pods.go:89] "kube-ingress-dns-minikube" [315f956b-f0bb-433b-9cc2-55581bcebdd4] Pending / Ready:ContainersNotReady (containers with unready status: [minikube-ingress-dns]) / ContainersReady:ContainersNotReady (containers with unready status: [minikube-ingress-dns])
I0917 16:57:03.481089 20150 system_pods.go:89] "kube-proxy-9xj99" [94e38d82-f813-4c73-ad0f-2b1d5bfd1a97] Running
I0917 16:57:03.481094 20150 system_pods.go:89] "kube-scheduler-addons-163060" [1a1eaf4b-4adc-4a9f-bcf3-2f8a56738f62] Running
I0917 16:57:03.481102 20150 system_pods.go:89] "metrics-server-84c5f94fbc-2f2f2" [03a25efb-5c8d-4637-b228-6bb67ccb601f] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0917 16:57:03.481108 20150 system_pods.go:89] "nvidia-device-plugin-daemonset-fvg2d" [69980d79-6040-46a6-92e4-f154f528e261] Pending / Ready:ContainersNotReady (containers with unready status: [nvidia-device-plugin-ctr]) / ContainersReady:ContainersNotReady (containers with unready status: [nvidia-device-plugin-ctr])
I0917 16:57:03.481125 20150 system_pods.go:89] "registry-66c9cd494c-xnftt" [87171e43-6b56-423a-ac20-6b46a3583197] Pending / Ready:ContainersNotReady (containers with unready status: [registry]) / ContainersReady:ContainersNotReady (containers with unready status: [registry])
I0917 16:57:03.481133 20150 system_pods.go:89] "registry-proxy-9ztsk" [de43c7a6-1992-4444-969d-d41949e06cdb] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0917 16:57:03.481138 20150 system_pods.go:89] "snapshot-controller-56fcc65765-fzkt2" [db3a7532-a6e9-481f-9449-e5e5f81fb4db] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0917 16:57:03.481147 20150 system_pods.go:89] "snapshot-controller-56fcc65765-t7gvk" [6e3f0826-ee2e-4a50-9797-2c8c8954b6a8] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0917 16:57:03.481155 20150 system_pods.go:89] "storage-provisioner" [bbad410c-6bb7-404e-bfd5-cb7d0e8f806c] Running
I0917 16:57:03.481162 20150 system_pods.go:89] "tiller-deploy-b48cc5f79-qd92x" [6de636f4-5713-4439-9d76-756777a66ef2] Pending / Ready:ContainersNotReady (containers with unready status: [tiller]) / ContainersReady:ContainersNotReady (containers with unready status: [tiller])
I0917 16:57:03.481168 20150 system_pods.go:126] duration metric: took 11.428374ms to wait for k8s-apps to be running ...
I0917 16:57:03.481177 20150 system_svc.go:44] waiting for kubelet service to be running ....
I0917 16:57:03.481215 20150 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0917 16:57:03.494547 20150 system_svc.go:56] duration metric: took 13.362249ms WaitForService to wait for kubelet
I0917 16:57:03.494572 20150 kubeadm.go:582] duration metric: took 15.135006283s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0917 16:57:03.494592 20150 node_conditions.go:102] verifying NodePressure condition ...
I0917 16:57:03.497623 20150 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0917 16:57:03.497657 20150 node_conditions.go:123] node cpu capacity is 8
I0917 16:57:03.497672 20150 node_conditions.go:105] duration metric: took 3.074367ms to run NodePressure ...
I0917 16:57:03.497686 20150 start.go:241] waiting for startup goroutines ...
I0917 16:57:03.664820 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:03.964200 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:03.964640 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:04.164643 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:04.464649 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:04.464884 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:04.664206 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:04.964808 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:04.964971 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:05.164559 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:05.464466 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:05.465556 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:05.663902 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:05.965020 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:05.965359 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:06.164234 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:06.465288 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:06.465637 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:06.663803 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:06.964941 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:06.965408 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:07.164286 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:07.464991 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:07.465939 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:07.664542 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:07.964303 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:07.964373 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:08.165598 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:08.463971 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:08.464252 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:08.663468 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:08.964405 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:08.964976 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:09.164239 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:09.465338 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:09.465532 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:09.663654 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:09.964213 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:09.964530 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:10.163564 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:10.465044 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:10.465231 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:10.663045 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:10.964019 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:10.964243 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:11.164544 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:11.464895 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:11.465260 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:11.664171 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:11.964755 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:11.966233 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:12.163953 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:12.464959 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:12.466455 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:12.664166 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:12.964029 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:12.964351 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:13.163756 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:13.464211 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:13.464319 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:13.663919 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:13.965040 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:13.965278 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:14.163793 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:14.464800 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:14.464901 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:14.663937 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:14.963975 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:14.964105 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:15.163126 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:15.465034 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:15.465640 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:15.664225 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:15.964599 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:15.965263 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:16.164691 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:16.465011 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:16.465394 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:16.664729 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:16.964873 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:16.965102 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:17.164441 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:17.463869 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:17.464374 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:17.663232 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:17.964155 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:17.965729 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:18.164193 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:18.464826 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:18.465055 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:18.665062 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:18.964257 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:18.964718 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:19.164092 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:19.464412 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:19.465154 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:19.663870 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:19.964088 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:19.964244 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:20.163807 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:20.464528 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:20.464734 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:20.663862 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:20.964172 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:20.964724 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:21.163037 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:21.463909 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:21.464247 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:21.663991 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:21.964667 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:21.964814 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:22.165154 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:22.464490 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:22.465616 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:22.663455 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:22.964368 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:22.964629 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:23.163574 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:23.464542 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:23.464673 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:23.663843 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:23.964417 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:23.965017 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:24.164544 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:24.464044 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:24.464390 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:24.663701 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:24.964882 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:24.965265 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:25.164525 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:25.464073 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:25.465382 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:25.663913 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:25.965522 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:25.965918 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:26.163718 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:26.464374 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:26.464888 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:26.664659 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:26.964176 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0917 16:57:26.964568 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:27.163759 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:27.464637 20150 kapi.go:107] duration metric: took 27.003377641s to wait for kubernetes.io/minikube-addons=registry ...
I0917 16:57:27.466002 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:27.663874 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:27.965336 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:28.163248 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:28.465073 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:28.663967 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:28.964343 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:29.163137 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:29.464667 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:29.663554 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:29.965194 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:30.164069 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:30.464950 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:30.664040 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:30.965486 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:31.163804 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:31.465692 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:31.663374 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:31.964359 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:32.163317 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:32.465812 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:32.664076 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:32.965154 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:33.164707 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:33.465098 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:33.665134 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:33.964876 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:34.163914 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:34.466326 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:34.664758 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:34.965074 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:35.164860 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:35.464952 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:35.664662 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:35.964857 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:36.164475 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:36.465417 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:36.665414 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:36.964839 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:37.164055 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:37.464759 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:37.664153 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:37.964605 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:38.163148 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:38.465747 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:38.664245 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:38.964934 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:39.164537 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:39.549800 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:39.663883 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:39.965733 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:40.164177 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:40.465045 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:40.664158 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:40.964624 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:41.163080 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:41.464689 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:41.663790 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:41.964758 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:42.164079 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:42.465301 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:42.664307 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:42.965186 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:43.163944 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:43.465551 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:43.663837 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:43.964072 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:44.165810 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:44.465601 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:44.663305 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:44.964561 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:45.163360 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:45.465349 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:45.662939 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:45.964520 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:46.164021 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:46.465641 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:46.665455 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:46.965148 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:47.164326 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:47.464429 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:47.663286 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:47.964881 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:48.163752 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:48.465770 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:48.664249 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:48.964630 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:49.164176 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:49.464595 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:49.664342 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:49.965844 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:50.163998 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:50.465278 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:50.664321 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:50.965260 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:51.162892 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:51.465141 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:51.664371 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:51.964627 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:52.163739 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:52.464681 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:52.663815 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:52.964752 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:53.164265 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:53.465112 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:53.664395 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:53.964518 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:54.164203 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:54.465615 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:54.664383 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:54.964864 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:55.163799 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:55.465086 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:55.664440 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:55.965144 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:56.163924 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:56.465327 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:56.664825 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:56.964931 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:57.163733 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:57.465099 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:57.664809 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:57.965028 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:58.164390 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0917 16:57:58.465602 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:58.663583 20150 kapi.go:107] duration metric: took 57.003993376s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0917 16:57:58.964775 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:59.464638 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:57:59.965231 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:00.464763 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:00.964754 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:01.464656 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:01.965042 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:02.465391 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:02.965368 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:03.465589 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:03.964782 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:04.465506 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:04.965776 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:05.465064 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:05.991596 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:06.465860 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:06.964213 20150 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0917 16:58:07.468751 20150 kapi.go:107] duration metric: took 1m7.007491906s to wait for app.kubernetes.io/name=ingress-nginx ...
I0917 16:58:25.460816 20150 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0917 16:58:25.460842 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:25.960698 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:26.460712 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:26.960720 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:27.460770 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:27.960566 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:28.460632 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:28.960825 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:29.461201 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:29.960822 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:30.460697 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:30.960591 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:31.460672 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:31.960319 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:32.461267 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:32.961478 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:33.461724 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:33.960560 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:34.460286 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:34.961396 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:35.460632 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:35.961493 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:36.461625 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:36.960987 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:37.460708 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:37.960715 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:38.460457 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:38.961534 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:39.460288 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:39.961234 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:40.461294 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:40.961562 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:41.461542 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:41.960845 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:42.460303 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:42.961321 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:43.460963 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:43.960840 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:44.461054 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:44.961006 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:45.461576 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:45.961661 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:46.461348 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:46.961524 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:47.461129 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:47.961139 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:48.460976 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:48.961255 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:49.461037 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:49.960823 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:50.460459 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:50.961587 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:51.460793 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:51.961019 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:52.460960 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:52.960789 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:53.461165 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:53.960934 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:54.460679 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:54.961707 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:55.461035 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:55.961088 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:56.461124 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:56.961213 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:57.461433 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:57.961297 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:58.461350 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:58.960772 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:59.460489 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:58:59.961407 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:00.461644 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:00.960746 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:01.460904 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:01.960845 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:02.460532 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:02.961669 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:03.460431 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:03.961258 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:04.461052 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:04.960837 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:05.461682 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:05.960292 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:06.460848 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:06.961358 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:07.461308 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:07.961160 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:08.461092 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:08.960560 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:09.460661 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:09.960262 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:10.461635 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:10.961445 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:11.461670 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:11.960752 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:12.460567 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:12.961735 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:13.460526 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:13.961318 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:14.461176 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:14.961225 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:15.461642 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:15.961304 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:16.460998 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:16.961479 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:17.461207 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:17.960984 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:18.460859 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:18.960878 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:19.460647 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:19.961598 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:20.460825 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:20.961829 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:21.460850 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:21.960865 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:22.460324 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:22.961532 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:23.461267 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:23.961386 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:24.461102 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:24.960510 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:25.460640 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:25.960487 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:26.461363 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:26.961394 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:27.461017 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:27.960767 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:28.460609 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:28.961767 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:29.460663 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:29.960862 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:30.460835 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:30.960638 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:31.461822 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:31.960730 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:32.460621 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:32.960757 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:33.460391 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:33.961356 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:34.461563 20150 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0917 16:59:34.960617 20150 kapi.go:107] duration metric: took 2m31.502971388s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0917 16:59:34.962322 20150 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-163060 cluster.
I0917 16:59:34.963591 20150 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0917 16:59:34.964898 20150 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0917 16:59:34.966347 20150 out.go:177] * Enabled addons: nvidia-device-plugin, storage-provisioner, default-storageclass, cloud-spanner, volcano, helm-tiller, ingress-dns, metrics-server, inspektor-gadget, yakd, storage-provisioner-rancher, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0917 16:59:34.967591 20150 addons.go:510] duration metric: took 2m46.608002975s for enable addons: enabled=[nvidia-device-plugin storage-provisioner default-storageclass cloud-spanner volcano helm-tiller ingress-dns metrics-server inspektor-gadget yakd storage-provisioner-rancher volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0917 16:59:34.967631 20150 start.go:246] waiting for cluster config update ...
I0917 16:59:34.967655 20150 start.go:255] writing updated cluster config ...
I0917 16:59:34.967917 20150 ssh_runner.go:195] Run: rm -f paused
I0917 16:59:35.017291 20150 start.go:600] kubectl: 1.31.1, cluster: 1.31.1 (minor skew: 0)
I0917 16:59:35.019095 20150 out.go:177] * Done! kubectl is now configured to use "addons-163060" cluster and "default" namespace by default
==> Docker <==
Sep 17 17:09:12 addons-163060 dockerd[1334]: time="2024-09-17T17:09:12.979865475Z" level=info msg="ignoring event" container=159b1218758378d2a78304e06b1a9f9cda65a9a48611fda7139632437cad6002 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:13 addons-163060 dockerd[1334]: time="2024-09-17T17:09:13.023483710Z" level=info msg="ignoring event" container=bbe0a8b5000e709ac317055de4fd6f66ec767f9084b32c9534aa0e95e1c30015 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:13 addons-163060 dockerd[1334]: time="2024-09-17T17:09:13.060725626Z" level=info msg="ignoring event" container=a330378b13c4d8a8286de11b872dfcfd31f72c382c77b9191b078c2cee380d0f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:14 addons-163060 dockerd[1334]: time="2024-09-17T17:09:14.201872175Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=c60bf6b9218da6012ad5a301ae71750169dac8a9f5778f37adf7d0947d822e5b
Sep 17 17:09:14 addons-163060 dockerd[1334]: time="2024-09-17T17:09:14.256313376Z" level=info msg="ignoring event" container=c60bf6b9218da6012ad5a301ae71750169dac8a9f5778f37adf7d0947d822e5b module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:14 addons-163060 dockerd[1334]: time="2024-09-17T17:09:14.376827278Z" level=info msg="ignoring event" container=0e95febb812b67559c1197d1284a23819910087faaf9c8294ea928215ee13cf5 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:15 addons-163060 dockerd[1334]: time="2024-09-17T17:09:15.592676037Z" level=info msg="Container failed to exit within 30s of signal 15 - using the force" container=501703641d76c4630fd3aa356bfc4c081648e5902586e7031b92537dcba107a3
Sep 17 17:09:15 addons-163060 dockerd[1334]: time="2024-09-17T17:09:15.613029862Z" level=info msg="ignoring event" container=501703641d76c4630fd3aa356bfc4c081648e5902586e7031b92537dcba107a3 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:15 addons-163060 dockerd[1334]: time="2024-09-17T17:09:15.748515162Z" level=info msg="ignoring event" container=ddfe6cdb3b5d0ad00423f7347a01b1020f3ba4cfd250e231e0c0fb37b800aeb2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:18 addons-163060 dockerd[1334]: time="2024-09-17T17:09:18.975853275Z" level=info msg="ignoring event" container=7894e13279ca91814cdaf3e5682ba1390780bcd4f1a0856788bddb3b2d6959ef module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:18 addons-163060 dockerd[1334]: time="2024-09-17T17:09:18.975900206Z" level=info msg="ignoring event" container=a5ecb62068e2e8adfdc9b1853c1c10d1b70a0270a42009134763948a1ca1d78f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:19 addons-163060 dockerd[1334]: time="2024-09-17T17:09:19.142914763Z" level=info msg="ignoring event" container=724666754cb92e31aba0afac8138c8cb188158161e05c52a7edc3854418272f9 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:19 addons-163060 dockerd[1334]: time="2024-09-17T17:09:19.172527770Z" level=info msg="ignoring event" container=8416d5e6bce98625407693d614211d8958c9e5d6e13f60c967e3ea3069e47e6c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:23 addons-163060 cri-dockerd[1599]: time="2024-09-17T17:09:23Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/d282c8307976b4d203712fcd0ea263f9b1d1c72cb05b34225d4e93d26ec3b6a8/resolv.conf as [nameserver 10.96.0.10 search kube-system.svc.cluster.local svc.cluster.local cluster.local europe-west4-a.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 17 17:09:26 addons-163060 cri-dockerd[1599]: time="2024-09-17T17:09:26Z" level=info msg="Stop pulling image docker.io/alpine/helm:2.16.3: Status: Downloaded newer image for alpine/helm:2.16.3"
Sep 17 17:09:26 addons-163060 dockerd[1334]: time="2024-09-17T17:09:26.308307043Z" level=info msg="ignoring event" container=426ff9e479ae24f94ec46f9c216265c31a0751314e417bba4df36744c118f5db module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:26 addons-163060 dockerd[1334]: time="2024-09-17T17:09:26.320856652Z" level=warning msg="failed to close stdin: NotFound: task 426ff9e479ae24f94ec46f9c216265c31a0751314e417bba4df36744c118f5db not found: not found"
Sep 17 17:09:28 addons-163060 dockerd[1334]: time="2024-09-17T17:09:28.118484130Z" level=info msg="ignoring event" container=d282c8307976b4d203712fcd0ea263f9b1d1c72cb05b34225d4e93d26ec3b6a8 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:28 addons-163060 dockerd[1334]: time="2024-09-17T17:09:28.231775378Z" level=info msg="ignoring event" container=be4db51af727a53a81ef4fbba1a26aac753218657cae7bee83736597c6fcab61 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:28 addons-163060 dockerd[1334]: time="2024-09-17T17:09:28.749468631Z" level=info msg="ignoring event" container=60e2ac8735b6f8db5223cfacbbefed38710c9c05ecf7486b442742195fb409a2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:28 addons-163060 dockerd[1334]: time="2024-09-17T17:09:28.862531501Z" level=info msg="ignoring event" container=8169aec9acef397d81986c1acebc4dc6db221b7003eee2edc865c63c89116960 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:28 addons-163060 dockerd[1334]: time="2024-09-17T17:09:28.964166973Z" level=info msg="ignoring event" container=7dee2065d2703369f696c55170e5055085fe05d7ac3ddee2e865180ecd1bb543 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:29 addons-163060 dockerd[1334]: time="2024-09-17T17:09:29.006289989Z" level=info msg="ignoring event" container=7b882d97f55e803a99df705046e9441d4a0bcf524b38ad4055d314d951df4e82 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:29 addons-163060 dockerd[1334]: time="2024-09-17T17:09:29.130776746Z" level=info msg="ignoring event" container=9c6e8e42c9d4621ab18d22411934a0afd5c6c26daf5c0d366f706adbd17f6ce5 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 17 17:09:29 addons-163060 dockerd[1334]: time="2024-09-17T17:09:29.276544381Z" level=info msg="ignoring event" container=72c0d7bb79141e77f13b705788e2dd8074781bcb99c608c0c75840b47a870ff9 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
8e510f831d2fc kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 18 seconds ago Running hello-world-app 0 e7653e88a50e5 hello-world-app-55bf9c44b4-g76gh
d21c6a4f836e8 nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf 28 seconds ago Running nginx 0 7f7b5a80a1cff nginx
46f177b415583 a416a98b71e22 44 seconds ago Exited helper-pod 0 aa8a938484c90 helper-pod-delete-pvc-6b40e24e-ff27-49e1-a0af-4a3320a2542e
4b0d6e930c47e busybox@sha256:c230832bd3b0be59a6c47ed64294f9ce71e91b327957920b6929a0caa8353140 47 seconds ago Exited busybox 0 29cd8dd943ecc test-local-path
8414040904fe8 gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 44dfe73f155c8 gcp-auth-89d5ffd79-hkv65
26b380342ba16 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited patch 0 a874b2ac10db7 ingress-nginx-admission-patch-9tgv9
08c9200d91012 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 6f81e5d9b5303 ingress-nginx-admission-create-x4jwn
9c6e8e42c9d46 gcr.io/k8s-minikube/kube-registry-proxy@sha256:b3fa0b2df8737fdb85ad5918a7e2652527463e357afff83a5e5bb966bcedc367 12 minutes ago Exited registry-proxy 0 72c0d7bb79141 registry-proxy-9ztsk
8a3331eedf321 6e38f40d628db 12 minutes ago Running storage-provisioner 0 40046a4413d03 storage-provisioner
2ecadef50c501 c69fa2e9cbf5f 12 minutes ago Running coredns 0 72c1b340f2c37 coredns-7c65d6cfc9-f8spg
793685e5c743a 60c005f310ff3 12 minutes ago Running kube-proxy 0 3559fb5338b2d kube-proxy-9xj99
831e94cfe3cec 175ffd71cce3d 12 minutes ago Running kube-controller-manager 0 1e33705e78b5c kube-controller-manager-addons-163060
4dc9603c31dc9 6bab7719df100 12 minutes ago Running kube-apiserver 0 9fb26e25b0543 kube-apiserver-addons-163060
70d4b9a9a5001 9aa1fad941575 12 minutes ago Running kube-scheduler 0 37bf23506c06a kube-scheduler-addons-163060
93ab0a3e97510 2e96e5913fc06 12 minutes ago Running etcd 0 d856cda90c09f etcd-addons-163060
==> coredns [2ecadef50c50] <==
[INFO] 10.244.0.7:50215 - 18049 "AAAA IN registry.kube-system.svc.cluster.local.svc.cluster.local. udp 74 false 512" NXDOMAIN qr,aa,rd 167 0.00015971s
[INFO] 10.244.0.7:55659 - 54909 "AAAA IN registry.kube-system.svc.cluster.local.cluster.local. udp 70 false 512" NXDOMAIN qr,aa,rd 163 0.000071428s
[INFO] 10.244.0.7:55659 - 41599 "A IN registry.kube-system.svc.cluster.local.cluster.local. udp 70 false 512" NXDOMAIN qr,aa,rd 163 0.000100948s
[INFO] 10.244.0.7:37046 - 44366 "A IN registry.kube-system.svc.cluster.local.europe-west4-a.c.k8s-minikube.internal. udp 95 false 512" NXDOMAIN qr,rd,ra 95 0.003810307s
[INFO] 10.244.0.7:37046 - 44619 "AAAA IN registry.kube-system.svc.cluster.local.europe-west4-a.c.k8s-minikube.internal. udp 95 false 512" NXDOMAIN qr,rd,ra 95 0.007955276s
[INFO] 10.244.0.7:42126 - 32729 "AAAA IN registry.kube-system.svc.cluster.local.c.k8s-minikube.internal. udp 80 false 512" NXDOMAIN qr,rd,ra 80 0.005689677s
[INFO] 10.244.0.7:42126 - 55975 "A IN registry.kube-system.svc.cluster.local.c.k8s-minikube.internal. udp 80 false 512" NXDOMAIN qr,rd,ra 80 0.007592471s
[INFO] 10.244.0.7:45729 - 47649 "A IN registry.kube-system.svc.cluster.local.google.internal. udp 72 false 512" NXDOMAIN qr,rd,ra 72 0.004609095s
[INFO] 10.244.0.7:45729 - 41532 "AAAA IN registry.kube-system.svc.cluster.local.google.internal. udp 72 false 512" NXDOMAIN qr,rd,ra 72 0.005486004s
[INFO] 10.244.0.7:34008 - 60102 "AAAA IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 149 0.00005795s
[INFO] 10.244.0.7:34008 - 2244 "A IN registry.kube-system.svc.cluster.local. udp 56 false 512" NOERROR qr,aa,rd 110 0.000116305s
[INFO] 10.244.0.26:37576 - 31457 "AAAA IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000325169s
[INFO] 10.244.0.26:34152 - 28088 "A IN storage.googleapis.com.gcp-auth.svc.cluster.local. udp 78 false 1232" NXDOMAIN qr,aa,rd 160 0.000427655s
[INFO] 10.244.0.26:57272 - 35175 "A IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000097913s
[INFO] 10.244.0.26:36498 - 2414 "AAAA IN storage.googleapis.com.svc.cluster.local. udp 69 false 1232" NXDOMAIN qr,aa,rd 151 0.000131391s
[INFO] 10.244.0.26:57583 - 3586 "AAAA IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000082477s
[INFO] 10.244.0.26:52040 - 10586 "A IN storage.googleapis.com.cluster.local. udp 65 false 1232" NXDOMAIN qr,aa,rd 147 0.000127909s
[INFO] 10.244.0.26:54644 - 21647 "AAAA IN storage.googleapis.com.europe-west4-a.c.k8s-minikube.internal. udp 90 false 1232" NXDOMAIN qr,rd,ra 79 0.007767534s
[INFO] 10.244.0.26:46035 - 63590 "A IN storage.googleapis.com.europe-west4-a.c.k8s-minikube.internal. udp 90 false 1232" NXDOMAIN qr,rd,ra 79 0.008630181s
[INFO] 10.244.0.26:43306 - 16437 "AAAA IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.007714913s
[INFO] 10.244.0.26:56839 - 38082 "A IN storage.googleapis.com.c.k8s-minikube.internal. udp 75 false 1232" NXDOMAIN qr,rd,ra 64 0.008665153s
[INFO] 10.244.0.26:42970 - 61231 "AAAA IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.006950152s
[INFO] 10.244.0.26:34126 - 43699 "A IN storage.googleapis.com.google.internal. udp 67 false 1232" NXDOMAIN qr,rd,ra 56 0.007022116s
[INFO] 10.244.0.26:57758 - 49619 "AAAA IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 240 0.003077072s
[INFO] 10.244.0.26:57649 - 24889 "A IN storage.googleapis.com. udp 51 false 1232" NOERROR qr,rd,ra 496 0.00320886s
==> describe nodes <==
Name: addons-163060
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-163060
kubernetes.io/os=linux
minikube.k8s.io/commit=825de77780746e57a7948604e1eea9da920a46ce
minikube.k8s.io/name=addons-163060
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_17T16_56_43_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-163060
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Tue, 17 Sep 2024 16:56:40 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-163060
AcquireTime: <unset>
RenewTime: Tue, 17 Sep 2024 17:09:28 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Tue, 17 Sep 2024 17:09:19 +0000 Tue, 17 Sep 2024 16:56:38 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Tue, 17 Sep 2024 17:09:19 +0000 Tue, 17 Sep 2024 16:56:38 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Tue, 17 Sep 2024 17:09:19 +0000 Tue, 17 Sep 2024 16:56:38 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Tue, 17 Sep 2024 17:09:19 +0000 Tue, 17 Sep 2024 16:56:40 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-163060
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859320Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859320Ki
pods: 110
System Info:
Machine ID: e6f018d39a5e4007b19afa3ae802fa60
System UUID: 0246bf5a-cd72-48db-a092-567c28e61886
Boot ID: 72a5ac5e-36f4-46e5-9bdc-b96891ef9823
Kernel Version: 5.15.0-1069-gcp
OS Image: Ubuntu 22.04.5 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.2.1
Kubelet Version: v1.31.1
Kube-Proxy Version: v1.31.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (11 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m14s
default hello-world-app-55bf9c44b4-g76gh 0 (0%) 0 (0%) 0 (0%) 0 (0%) 20s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 32s
gcp-auth gcp-auth-89d5ffd79-hkv65 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
kube-system coredns-7c65d6cfc9-f8spg 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-163060 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-163060 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-163060 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-9xj99 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-163060 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-163060 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-163060 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-163060 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-163060 event: Registered Node addons-163060 in Controller
==> dmesg <==
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 26 eb fd f2 7e f1 08 06
[ +2.746902] IPv4: martian source 10.244.0.1 from 10.244.0.17, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff a2 9f 8c c1 dd 2b 08 06
[ +2.611911] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 06 44 db 35 1d a3 08 06
[ +6.069320] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff d2 f5 5d e2 dc f7 08 06
[ +0.244288] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff fa 77 b5 45 be f9 08 06
[ +0.019454] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff d6 e4 d4 92 e1 7d 08 06
[Sep17 16:58] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff c2 c5 90 0d e1 85 08 06
[Sep17 16:59] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff d6 97 ce f8 5f 17 08 06
[ +0.023828] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 3a 36 1e ef 35 bf 08 06
[ +27.507234] IPv4: martian source 10.244.0.1 from 10.244.0.26, on dev eth0
[ +0.000018] ll header: 00000000: ff ff ff ff ff ff 3a 7e 88 27 60 c4 08 06
[ +0.000576] IPv4: martian source 10.244.0.26 from 10.244.0.3, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff ee 1e c7 c9 38 61 08 06
[Sep17 17:09] IPv4: martian source 10.244.0.35 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff c2 c5 90 0d e1 85 08 06
[ +17.308214] IPv4: martian source 10.244.0.1 from 10.244.0.38, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 6a 3b 5f b3 33 e6 08 06
==> etcd [93ab0a3e9751] <==
{"level":"info","ts":"2024-09-17T16:56:38.247802Z","caller":"embed/etcd.go:571","msg":"cmux::serve","address":"192.168.49.2:2380"}
{"level":"info","ts":"2024-09-17T16:56:38.877302Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc is starting a new election at term 1"}
{"level":"info","ts":"2024-09-17T16:56:38.877344Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became pre-candidate at term 1"}
{"level":"info","ts":"2024-09-17T16:56:38.877365Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgPreVoteResp from aec36adc501070cc at term 1"}
{"level":"info","ts":"2024-09-17T16:56:38.877390Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became candidate at term 2"}
{"level":"info","ts":"2024-09-17T16:56:38.877400Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-17T16:56:38.877416Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-09-17T16:56:38.877430Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-17T16:56:38.878378Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-163060 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-09-17T16:56:38.878392Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-17T16:56:38.878442Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-17T16:56:38.878474Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-17T16:56:38.878614Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-17T16:56:38.878652Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-17T16:56:38.879157Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-17T16:56:38.879413Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-17T16:56:38.879444Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-17T16:56:38.879659Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-17T16:56:38.879662Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-17T16:56:38.880689Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-17T16:56:38.880790Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-09-17T16:56:58.598768Z","caller":"traceutil/trace.go:171","msg":"trace[1057188115] transaction","detail":"{read_only:false; response_revision:758; number_of_response:1; }","duration":"132.474288ms","start":"2024-09-17T16:56:58.466274Z","end":"2024-09-17T16:56:58.598749Z","steps":["trace[1057188115] 'process raft request' (duration: 111.016342ms)","trace[1057188115] 'compare' (duration: 21.262115ms)"],"step_count":2}
{"level":"info","ts":"2024-09-17T17:06:38.898812Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1875}
{"level":"info","ts":"2024-09-17T17:06:38.925702Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1875,"took":"26.299488ms","hash":2471208411,"current-db-size-bytes":9146368,"current-db-size":"9.1 MB","current-db-size-in-use-bytes":4919296,"current-db-size-in-use":"4.9 MB"}
{"level":"info","ts":"2024-09-17T17:06:38.925746Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":2471208411,"revision":1875,"compact-revision":-1}
==> gcp-auth [8414040904fe] <==
2024/09/17 17:00:15 Ready to write response ...
2024/09/17 17:08:28 Ready to marshal response ...
2024/09/17 17:08:28 Ready to write response ...
2024/09/17 17:08:33 Ready to marshal response ...
2024/09/17 17:08:33 Ready to write response ...
2024/09/17 17:08:34 Ready to marshal response ...
2024/09/17 17:08:34 Ready to write response ...
2024/09/17 17:08:34 Ready to marshal response ...
2024/09/17 17:08:34 Ready to write response ...
2024/09/17 17:08:40 Ready to marshal response ...
2024/09/17 17:08:40 Ready to write response ...
2024/09/17 17:08:40 Ready to marshal response ...
2024/09/17 17:08:40 Ready to write response ...
2024/09/17 17:08:40 Ready to marshal response ...
2024/09/17 17:08:40 Ready to write response ...
2024/09/17 17:08:44 Ready to marshal response ...
2024/09/17 17:08:44 Ready to write response ...
2024/09/17 17:08:57 Ready to marshal response ...
2024/09/17 17:08:57 Ready to write response ...
2024/09/17 17:09:02 Ready to marshal response ...
2024/09/17 17:09:02 Ready to write response ...
2024/09/17 17:09:09 Ready to marshal response ...
2024/09/17 17:09:09 Ready to write response ...
2024/09/17 17:09:23 Ready to marshal response ...
2024/09/17 17:09:23 Ready to write response ...
==> kernel <==
17:09:29 up 51 min, 0 users, load average: 0.13, 0.30, 0.34
Linux addons-163060 5.15.0-1069-gcp #77~20.04.1-Ubuntu SMP Sun Sep 1 19:39:16 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.5 LTS"
==> kube-apiserver [4dc9603c31dc] <==
W0917 17:00:07.951820 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0917 17:00:08.075079 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0917 17:00:08.461289 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0917 17:08:23.387730 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0917 17:08:24.404465 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0917 17:08:33.491569 1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
I0917 17:08:40.595998 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.101.151.56"}
I0917 17:08:41.315137 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0917 17:08:57.441514 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0917 17:08:57.659595 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.107.204.62"}
E0917 17:09:00.981873 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
I0917 17:09:09.143474 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.111.19.10"}
I0917 17:09:18.829020 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0917 17:09:18.829069 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0917 17:09:18.841694 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0917 17:09:18.841751 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0917 17:09:18.842623 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0917 17:09:18.842670 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0917 17:09:18.853749 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0917 17:09:18.853800 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0917 17:09:18.863193 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0917 17:09:18.863234 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0917 17:09:19.843348 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W0917 17:09:19.864241 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
W0917 17:09:19.873240 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
==> kube-controller-manager [831e94cfe3ce] <==
W0917 17:09:20.827664 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:20.827701 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 17:09:20.890655 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:20.890692 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 17:09:20.997443 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:20.997480 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0917 17:09:21.341529 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="ingress-nginx"
W0917 17:09:23.220170 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:23.220211 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 17:09:23.330268 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:23.330308 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 17:09:23.336615 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:23.336649 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 17:09:23.635908 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:23.635942 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 17:09:24.042630 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:24.042696 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 17:09:26.944453 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:26.944498 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 17:09:28.493896 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:28.493932 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0917 17:09:28.513863 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0917 17:09:28.513914 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0917 17:09:28.657051 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/tiller-deploy-b48cc5f79" duration="9.798µs"
I0917 17:09:28.780831 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="9.78µs"
==> kube-proxy [793685e5c743] <==
I0917 16:56:48.348771 1 server_linux.go:66] "Using iptables proxy"
I0917 16:56:48.565320 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0917 16:56:48.565396 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0917 16:56:48.766790 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0917 16:56:48.766855 1 server_linux.go:169] "Using iptables Proxier"
I0917 16:56:48.772854 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0917 16:56:48.773143 1 server.go:483] "Version info" version="v1.31.1"
I0917 16:56:48.773166 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0917 16:56:48.774683 1 config.go:199] "Starting service config controller"
I0917 16:56:48.774700 1 shared_informer.go:313] Waiting for caches to sync for service config
I0917 16:56:48.774731 1 config.go:105] "Starting endpoint slice config controller"
I0917 16:56:48.774736 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0917 16:56:48.774952 1 config.go:328] "Starting node config controller"
I0917 16:56:48.775061 1 shared_informer.go:313] Waiting for caches to sync for node config
I0917 16:56:48.947174 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0917 16:56:48.947247 1 shared_informer.go:320] Caches are synced for service config
I0917 16:56:48.947633 1 shared_informer.go:320] Caches are synced for node config
==> kube-scheduler [70d4b9a9a500] <==
W0917 16:56:40.448764 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0917 16:56:40.448781 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0917 16:56:40.448891 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0917 16:56:40.449624 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 16:56:40.448891 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0917 16:56:40.450747 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:kube-scheduler\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 16:56:40.448913 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0917 16:56:40.450789 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0917 16:56:40.448977 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0917 16:56:40.450818 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 16:56:40.448989 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0917 16:56:40.450856 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 16:56:40.449019 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0917 16:56:40.450882 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 16:56:41.272631 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0917 16:56:41.272668 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 16:56:41.281764 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0917 16:56:41.281795 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0917 16:56:41.403776 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0917 16:56:41.403812 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 16:56:41.413003 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0917 16:56:41.413045 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0917 16:56:41.551492 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0917 16:56:41.551543 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
I0917 16:56:41.866788 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 17 17:09:28 addons-163060 kubelet[2458]: I0917 17:09:28.302460 2458 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/bac1bb97-ad10-402c-bcc5-3417b7af8ed6-gcp-creds\") pod \"bac1bb97-ad10-402c-bcc5-3417b7af8ed6\" (UID: \"bac1bb97-ad10-402c-bcc5-3417b7af8ed6\") "
Sep 17 17:09:28 addons-163060 kubelet[2458]: I0917 17:09:28.302492 2458 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfhs4\" (UniqueName: \"kubernetes.io/projected/5ad8a734-ffc8-4d65-8883-944f32116156-kube-api-access-wfhs4\") pod \"5ad8a734-ffc8-4d65-8883-944f32116156\" (UID: \"5ad8a734-ffc8-4d65-8883-944f32116156\") "
Sep 17 17:09:28 addons-163060 kubelet[2458]: I0917 17:09:28.303102 2458 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bac1bb97-ad10-402c-bcc5-3417b7af8ed6-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "bac1bb97-ad10-402c-bcc5-3417b7af8ed6" (UID: "bac1bb97-ad10-402c-bcc5-3417b7af8ed6"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Sep 17 17:09:28 addons-163060 kubelet[2458]: I0917 17:09:28.350296 2458 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad8a734-ffc8-4d65-8883-944f32116156-kube-api-access-wfhs4" (OuterVolumeSpecName: "kube-api-access-wfhs4") pod "5ad8a734-ffc8-4d65-8883-944f32116156" (UID: "5ad8a734-ffc8-4d65-8883-944f32116156"). InnerVolumeSpecName "kube-api-access-wfhs4". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 17 17:09:28 addons-163060 kubelet[2458]: I0917 17:09:28.352561 2458 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bac1bb97-ad10-402c-bcc5-3417b7af8ed6-kube-api-access-4m7pf" (OuterVolumeSpecName: "kube-api-access-4m7pf") pod "bac1bb97-ad10-402c-bcc5-3417b7af8ed6" (UID: "bac1bb97-ad10-402c-bcc5-3417b7af8ed6"). InnerVolumeSpecName "kube-api-access-4m7pf". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 17 17:09:28 addons-163060 kubelet[2458]: I0917 17:09:28.403452 2458 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-4m7pf\" (UniqueName: \"kubernetes.io/projected/bac1bb97-ad10-402c-bcc5-3417b7af8ed6-kube-api-access-4m7pf\") on node \"addons-163060\" DevicePath \"\""
Sep 17 17:09:28 addons-163060 kubelet[2458]: I0917 17:09:28.403482 2458 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/bac1bb97-ad10-402c-bcc5-3417b7af8ed6-gcp-creds\") on node \"addons-163060\" DevicePath \"\""
Sep 17 17:09:28 addons-163060 kubelet[2458]: I0917 17:09:28.403490 2458 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-wfhs4\" (UniqueName: \"kubernetes.io/projected/5ad8a734-ffc8-4d65-8883-944f32116156-kube-api-access-wfhs4\") on node \"addons-163060\" DevicePath \"\""
Sep 17 17:09:28 addons-163060 kubelet[2458]: I0917 17:09:28.462463 2458 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="kube-system/registry-proxy-9ztsk" secret="" err="secret \"gcp-auth\" not found"
Sep 17 17:09:28 addons-163060 kubelet[2458]: I0917 17:09:28.471941 2458 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad8a734-ffc8-4d65-8883-944f32116156" path="/var/lib/kubelet/pods/5ad8a734-ffc8-4d65-8883-944f32116156/volumes"
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.027504 2458 scope.go:117] "RemoveContainer" containerID="426ff9e479ae24f94ec46f9c216265c31a0751314e417bba4df36744c118f5db"
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.042695 2458 scope.go:117] "RemoveContainer" containerID="8169aec9acef397d81986c1acebc4dc6db221b7003eee2edc865c63c89116960"
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.064544 2458 scope.go:117] "RemoveContainer" containerID="60e2ac8735b6f8db5223cfacbbefed38710c9c05ecf7486b442742195fb409a2"
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.082749 2458 scope.go:117] "RemoveContainer" containerID="60e2ac8735b6f8db5223cfacbbefed38710c9c05ecf7486b442742195fb409a2"
Sep 17 17:09:29 addons-163060 kubelet[2458]: E0917 17:09:29.083626 2458 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 60e2ac8735b6f8db5223cfacbbefed38710c9c05ecf7486b442742195fb409a2" containerID="60e2ac8735b6f8db5223cfacbbefed38710c9c05ecf7486b442742195fb409a2"
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.083668 2458 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"60e2ac8735b6f8db5223cfacbbefed38710c9c05ecf7486b442742195fb409a2"} err="failed to get container status \"60e2ac8735b6f8db5223cfacbbefed38710c9c05ecf7486b442742195fb409a2\": rpc error: code = Unknown desc = Error response from daemon: No such container: 60e2ac8735b6f8db5223cfacbbefed38710c9c05ecf7486b442742195fb409a2"
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.149980 2458 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scps5\" (UniqueName: \"kubernetes.io/projected/87171e43-6b56-423a-ac20-6b46a3583197-kube-api-access-scps5\") pod \"87171e43-6b56-423a-ac20-6b46a3583197\" (UID: \"87171e43-6b56-423a-ac20-6b46a3583197\") "
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.150045 2458 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsf2k\" (UniqueName: \"kubernetes.io/projected/6de636f4-5713-4439-9d76-756777a66ef2-kube-api-access-nsf2k\") pod \"6de636f4-5713-4439-9d76-756777a66ef2\" (UID: \"6de636f4-5713-4439-9d76-756777a66ef2\") "
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.151866 2458 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87171e43-6b56-423a-ac20-6b46a3583197-kube-api-access-scps5" (OuterVolumeSpecName: "kube-api-access-scps5") pod "87171e43-6b56-423a-ac20-6b46a3583197" (UID: "87171e43-6b56-423a-ac20-6b46a3583197"). InnerVolumeSpecName "kube-api-access-scps5". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.152032 2458 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6de636f4-5713-4439-9d76-756777a66ef2-kube-api-access-nsf2k" (OuterVolumeSpecName: "kube-api-access-nsf2k") pod "6de636f4-5713-4439-9d76-756777a66ef2" (UID: "6de636f4-5713-4439-9d76-756777a66ef2"). InnerVolumeSpecName "kube-api-access-nsf2k". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.250560 2458 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-nsf2k\" (UniqueName: \"kubernetes.io/projected/6de636f4-5713-4439-9d76-756777a66ef2-kube-api-access-nsf2k\") on node \"addons-163060\" DevicePath \"\""
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.250601 2458 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-scps5\" (UniqueName: \"kubernetes.io/projected/87171e43-6b56-423a-ac20-6b46a3583197-kube-api-access-scps5\") on node \"addons-163060\" DevicePath \"\""
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.452348 2458 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljhb9\" (UniqueName: \"kubernetes.io/projected/de43c7a6-1992-4444-969d-d41949e06cdb-kube-api-access-ljhb9\") pod \"de43c7a6-1992-4444-969d-d41949e06cdb\" (UID: \"de43c7a6-1992-4444-969d-d41949e06cdb\") "
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.454087 2458 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de43c7a6-1992-4444-969d-d41949e06cdb-kube-api-access-ljhb9" (OuterVolumeSpecName: "kube-api-access-ljhb9") pod "de43c7a6-1992-4444-969d-d41949e06cdb" (UID: "de43c7a6-1992-4444-969d-d41949e06cdb"). InnerVolumeSpecName "kube-api-access-ljhb9". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 17 17:09:29 addons-163060 kubelet[2458]: I0917 17:09:29.552867 2458 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-ljhb9\" (UniqueName: \"kubernetes.io/projected/de43c7a6-1992-4444-969d-d41949e06cdb-kube-api-access-ljhb9\") on node \"addons-163060\" DevicePath \"\""
==> storage-provisioner [8a3331eedf32] <==
I0917 16:56:55.870387 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0917 16:56:55.960966 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0917 16:56:55.961016 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0917 16:56:56.049106 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0917 16:56:56.049320 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-163060_546f4f54-f306-4def-a157-3e9073616e90!
I0917 16:56:56.052387 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"12e461c2-dc4f-4e57-8763-ba4b004ba039", APIVersion:"v1", ResourceVersion:"602", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-163060_546f4f54-f306-4def-a157-3e9073616e90 became leader
I0917 16:56:56.251069 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-163060_546f4f54-f306-4def-a157-3e9073616e90!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-163060 -n addons-163060
helpers_test.go:261: (dbg) Run: kubectl --context addons-163060 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-163060 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-163060 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-163060/192.168.49.2
Start Time: Tue, 17 Sep 2024 17:00:15 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.28
IPs:
IP: 10.244.0.28
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-qb4wj (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-qb4wj:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m15s default-scheduler Successfully assigned default/busybox to addons-163060
Normal Pulling 7m47s (x4 over 9m14s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m47s (x4 over 9m14s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m47s (x4 over 9m14s) kubelet Error: ErrImagePull
Warning Failed 7m31s (x6 over 9m14s) kubelet Error: ImagePullBackOff
Normal BackOff 4m9s (x20 over 9m14s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (72.63s)