=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:332: registry stabilized in 1.90331ms
addons_test.go:334: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-66c9cd494c-t9hlq" [48d3ee3d-2a77-4bbb-9492-d6cb8accc5e0] Running
addons_test.go:334: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.002552172s
addons_test.go:337: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-k8hwr" [8343911e-7ef0-4555-a6d5-acefa24b2ae3] Running
addons_test.go:337: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.003515147s
addons_test.go:342: (dbg) Run: kubectl --context addons-224201 delete po -l run=registry-test --now
addons_test.go:347: (dbg) Run: kubectl --context addons-224201 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:347: (dbg) Non-zero exit: kubectl --context addons-224201 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.07060553s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:349: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-224201 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:353: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:361: (dbg) Run: out/minikube-linux-amd64 -p addons-224201 ip
2024/09/15 06:42:56 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:390: (dbg) Run: out/minikube-linux-amd64 -p addons-224201 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-224201
helpers_test.go:235: (dbg) docker inspect addons-224201:
-- stdout --
[
{
"Id": "08abb29397a5ccdb31f96c113dd189cee1c4126d78eec2473f6764d708f8395b",
"Created": "2024-09-15T06:30:00.250941294Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 14800,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-15T06:30:00.37308508Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:20d492278eed428d119466f58713403332b5d2ac1db7c6863f797e2406f2b671",
"ResolvConfPath": "/var/lib/docker/containers/08abb29397a5ccdb31f96c113dd189cee1c4126d78eec2473f6764d708f8395b/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/08abb29397a5ccdb31f96c113dd189cee1c4126d78eec2473f6764d708f8395b/hostname",
"HostsPath": "/var/lib/docker/containers/08abb29397a5ccdb31f96c113dd189cee1c4126d78eec2473f6764d708f8395b/hosts",
"LogPath": "/var/lib/docker/containers/08abb29397a5ccdb31f96c113dd189cee1c4126d78eec2473f6764d708f8395b/08abb29397a5ccdb31f96c113dd189cee1c4126d78eec2473f6764d708f8395b-json.log",
"Name": "/addons-224201",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"/lib/modules:/lib/modules:ro",
"addons-224201:/var"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-224201",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/aabc48a535f768279fd9cbd80bef4b6377aa3ed4c0571492c5dbec725e8ee296-init/diff:/var/lib/docker/overlay2/f021d4ce5dac28fd395028054f759c8d42bf3696086cb2248e79dd0e572c19c4/diff",
"MergedDir": "/var/lib/docker/overlay2/aabc48a535f768279fd9cbd80bef4b6377aa3ed4c0571492c5dbec725e8ee296/merged",
"UpperDir": "/var/lib/docker/overlay2/aabc48a535f768279fd9cbd80bef4b6377aa3ed4c0571492c5dbec725e8ee296/diff",
"WorkDir": "/var/lib/docker/overlay2/aabc48a535f768279fd9cbd80bef4b6377aa3ed4c0571492c5dbec725e8ee296/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
},
{
"Type": "volume",
"Name": "addons-224201",
"Source": "/var/lib/docker/volumes/addons-224201/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
}
],
"Config": {
"Hostname": "addons-224201",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-224201",
"name.minikube.sigs.k8s.io": "addons-224201",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "1f101b1f41fc6a03ce7ac9e22513625968f749041fb0f39d95250efb37b02c21",
"SandboxKey": "/var/run/docker/netns/1f101b1f41fc",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-224201": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "667c34e40711b7c17a93c31eb604d8dec79f2ad036230b70b94f299ed7783dc3",
"EndpointID": "eb809e735d2e8624d7124779afa8610b82838112eb9a70c8e778fd8edd0aae84",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-224201",
"08abb29397a5"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-224201 -n addons-224201
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-224201 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-docker-946025 | download-docker-946025 | jenkins | v1.34.0 | 15 Sep 24 06:29 UTC | 15 Sep 24 06:29 UTC |
| start | --download-only -p | binary-mirror-451815 | jenkins | v1.34.0 | 15 Sep 24 06:29 UTC | |
| | binary-mirror-451815 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:41433 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-451815 | binary-mirror-451815 | jenkins | v1.34.0 | 15 Sep 24 06:29 UTC | 15 Sep 24 06:29 UTC |
| addons | enable dashboard -p | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:29 UTC | |
| | addons-224201 | | | | | |
| addons | disable dashboard -p | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:29 UTC | |
| | addons-224201 | | | | | |
| start | -p addons-224201 --wait=true | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:29 UTC | 15 Sep 24 06:33 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | addons-224201 addons disable | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:33 UTC | 15 Sep 24 06:33 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | addons-224201 addons disable | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:41 UTC | 15 Sep 24 06:41 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| ssh | addons-224201 ssh cat | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:41 UTC | 15 Sep 24 06:41 UTC |
| | /opt/local-path-provisioner/pvc-4c0e3716-a4e0-4e68-9650-d0af4b5f869d_default_test-pvc/file1 | | | | | |
| addons | addons-224201 addons disable | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:41 UTC | 15 Sep 24 06:41 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:41 UTC | 15 Sep 24 06:42 UTC |
| | -p addons-224201 | | | | | |
| addons | disable cloud-spanner -p | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | addons-224201 | | | | | |
| addons | addons-224201 addons | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | enable headlamp | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | -p addons-224201 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-224201 addons disable | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-224201 addons disable | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | addons-224201 | | | | | |
| addons | addons-224201 addons | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-224201 addons | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-224201 ssh curl -s | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-224201 ip | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| addons | addons-224201 addons disable | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-224201 addons disable | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| ip | addons-224201 ip | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| addons | addons-224201 addons disable | addons-224201 | jenkins | v1.34.0 | 15 Sep 24 06:42 UTC | 15 Sep 24 06:42 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/15 06:29:36
Running on machine: ubuntu-20-agent-14
Binary: Built with gc go1.23.0 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0915 06:29:36.967016 14042 out.go:345] Setting OutFile to fd 1 ...
I0915 06:29:36.967110 14042 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0915 06:29:36.967120 14042 out.go:358] Setting ErrFile to fd 2...
I0915 06:29:36.967124 14042 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0915 06:29:36.967298 14042 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19644-6048/.minikube/bin
I0915 06:29:36.967840 14042 out.go:352] Setting JSON to false
I0915 06:29:36.968624 14042 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":725,"bootTime":1726381052,"procs":172,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1068-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0915 06:29:36.968707 14042 start.go:139] virtualization: kvm guest
I0915 06:29:36.970551 14042 out.go:177] * [addons-224201] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0915 06:29:36.971774 14042 notify.go:220] Checking for updates...
I0915 06:29:36.971782 14042 out.go:177] - MINIKUBE_LOCATION=19644
I0915 06:29:36.972770 14042 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0915 06:29:36.973737 14042 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19644-6048/kubeconfig
I0915 06:29:36.974797 14042 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19644-6048/.minikube
I0915 06:29:36.975851 14042 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0915 06:29:36.976772 14042 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0915 06:29:36.977718 14042 driver.go:394] Setting default libvirt URI to qemu:///system
I0915 06:29:36.997343 14042 docker.go:123] docker version: linux-27.2.1:Docker Engine - Community
I0915 06:29:36.997411 14042 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0915 06:29:37.040715 14042 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-15 06:29:37.03280553 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86_
64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErr
ors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0915 06:29:37.040805 14042 docker.go:318] overlay module found
I0915 06:29:37.043107 14042 out.go:177] * Using the docker driver based on user configuration
I0915 06:29:37.044017 14042 start.go:297] selected driver: docker
I0915 06:29:37.044025 14042 start.go:901] validating driver "docker" against <nil>
I0915 06:29:37.044034 14042 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0915 06:29:37.044721 14042 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0915 06:29:37.090853 14042 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-15 06:29:37.082906411 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1068-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.1 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c Expected:7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c} RuncCommit:{ID:v1.1.14-0-g2c9f560 Expected:v1.1.14-0-g2c9f560} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0915 06:29:37.091038 14042 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0915 06:29:37.091268 14042 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0915 06:29:37.092558 14042 out.go:177] * Using Docker driver with root privileges
I0915 06:29:37.093792 14042 cni.go:84] Creating CNI manager for ""
I0915 06:29:37.093849 14042 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0915 06:29:37.093861 14042 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0915 06:29:37.093918 14042 start.go:340] cluster config:
{Name:addons-224201 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-224201 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0915 06:29:37.095456 14042 out.go:177] * Starting "addons-224201" primary control-plane node in "addons-224201" cluster
I0915 06:29:37.096540 14042 cache.go:121] Beginning downloading kic base image for docker with docker
I0915 06:29:37.097569 14042 out.go:177] * Pulling base image v0.0.45-1726358845-19644 ...
I0915 06:29:37.098539 14042 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0915 06:29:37.098568 14042 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 in local docker daemon
I0915 06:29:37.098578 14042 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19644-6048/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4
I0915 06:29:37.098586 14042 cache.go:56] Caching tarball of preloaded images
I0915 06:29:37.098706 14042 preload.go:172] Found /home/jenkins/minikube-integration/19644-6048/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0915 06:29:37.098720 14042 cache.go:59] Finished verifying existence of preloaded tar for v1.31.1 on docker
I0915 06:29:37.099045 14042 profile.go:143] Saving config to /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/config.json ...
I0915 06:29:37.099068 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/config.json: {Name:mkd174c094dbaeb9bbb5dc0bbe22077ebf72e14c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:29:37.113081 14042 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 to local cache
I0915 06:29:37.113166 14042 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 in local cache directory
I0915 06:29:37.113184 14042 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 in local cache directory, skipping pull
I0915 06:29:37.113191 14042 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 exists in cache, skipping pull
I0915 06:29:37.113207 14042 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 as a tarball
I0915 06:29:37.113214 14042 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 from local cache
I0915 06:29:48.788772 14042 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 from cached tarball
I0915 06:29:48.788805 14042 cache.go:194] Successfully downloaded all kic artifacts
I0915 06:29:48.788850 14042 start.go:360] acquireMachinesLock for addons-224201: {Name:mk3c25920ea60a5a39ec14c5bf3dc93343588a64 Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0915 06:29:48.788943 14042 start.go:364] duration metric: took 74.913µs to acquireMachinesLock for "addons-224201"
I0915 06:29:48.788968 14042 start.go:93] Provisioning new machine with config: &{Name:addons-224201 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-224201 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0915 06:29:48.789055 14042 start.go:125] createHost starting for "" (driver="docker")
I0915 06:29:48.790501 14042 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0915 06:29:48.790773 14042 start.go:159] libmachine.API.Create for "addons-224201" (driver="docker")
I0915 06:29:48.790808 14042 client.go:168] LocalClient.Create starting
I0915 06:29:48.790902 14042 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19644-6048/.minikube/certs/ca.pem
I0915 06:29:48.929680 14042 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19644-6048/.minikube/certs/cert.pem
I0915 06:29:49.132907 14042 cli_runner.go:164] Run: docker network inspect addons-224201 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0915 06:29:49.147880 14042 cli_runner.go:211] docker network inspect addons-224201 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0915 06:29:49.147953 14042 network_create.go:284] running [docker network inspect addons-224201] to gather additional debugging logs...
I0915 06:29:49.147971 14042 cli_runner.go:164] Run: docker network inspect addons-224201
W0915 06:29:49.161997 14042 cli_runner.go:211] docker network inspect addons-224201 returned with exit code 1
I0915 06:29:49.162019 14042 network_create.go:287] error running [docker network inspect addons-224201]: docker network inspect addons-224201: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-224201 not found
I0915 06:29:49.162033 14042 network_create.go:289] output of [docker network inspect addons-224201]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-224201 not found
** /stderr **
I0915 06:29:49.162106 14042 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0915 06:29:49.175656 14042 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc00039fa60}
I0915 06:29:49.175711 14042 network_create.go:124] attempt to create docker network addons-224201 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0915 06:29:49.175763 14042 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-224201 addons-224201
I0915 06:29:49.229182 14042 network_create.go:108] docker network addons-224201 192.168.49.0/24 created
I0915 06:29:49.229214 14042 kic.go:121] calculated static IP "192.168.49.2" for the "addons-224201" container
I0915 06:29:49.229268 14042 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0915 06:29:49.243060 14042 cli_runner.go:164] Run: docker volume create addons-224201 --label name.minikube.sigs.k8s.io=addons-224201 --label created_by.minikube.sigs.k8s.io=true
I0915 06:29:49.258670 14042 oci.go:103] Successfully created a docker volume addons-224201
I0915 06:29:49.258737 14042 cli_runner.go:164] Run: docker run --rm --name addons-224201-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-224201 --entrypoint /usr/bin/test -v addons-224201:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -d /var/lib
I0915 06:29:56.405305 14042 cli_runner.go:217] Completed: docker run --rm --name addons-224201-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-224201 --entrypoint /usr/bin/test -v addons-224201:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -d /var/lib: (7.146524759s)
I0915 06:29:56.405331 14042 oci.go:107] Successfully prepared a docker volume addons-224201
I0915 06:29:56.405356 14042 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0915 06:29:56.405375 14042 kic.go:194] Starting extracting preloaded images to volume ...
I0915 06:29:56.405431 14042 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19644-6048/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-224201:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -I lz4 -xf /preloaded.tar -C /extractDir
I0915 06:30:00.187102 14042 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19644-6048/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.1-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-224201:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 -I lz4 -xf /preloaded.tar -C /extractDir: (3.781637635s)
I0915 06:30:00.187126 14042 kic.go:203] duration metric: took 3.781749324s to extract preloaded images to volume ...
W0915 06:30:00.187243 14042 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0915 06:30:00.187377 14042 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0915 06:30:00.236483 14042 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-224201 --name addons-224201 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-224201 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-224201 --network addons-224201 --ip 192.168.49.2 --volume addons-224201:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0
I0915 06:30:00.546964 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Running}}
I0915 06:30:00.564798 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:00.582261 14042 cli_runner.go:164] Run: docker exec addons-224201 stat /var/lib/dpkg/alternatives/iptables
I0915 06:30:00.623578 14042 oci.go:144] the created container "addons-224201" has a running status.
I0915 06:30:00.623607 14042 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa...
I0915 06:30:00.811307 14042 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0915 06:30:00.835783 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:00.854803 14042 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0915 06:30:00.854834 14042 kic_runner.go:114] Args: [docker exec --privileged addons-224201 chown docker:docker /home/docker/.ssh/authorized_keys]
I0915 06:30:00.970485 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:00.990860 14042 machine.go:93] provisionDockerMachine start ...
I0915 06:30:00.990939 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:01.013313 14042 main.go:141] libmachine: Using SSH client type: native
I0915 06:30:01.013587 14042 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 06:30:01.013604 14042 main.go:141] libmachine: About to run SSH command:
hostname
I0915 06:30:01.177222 14042 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-224201
I0915 06:30:01.177250 14042 ubuntu.go:169] provisioning hostname "addons-224201"
I0915 06:30:01.177309 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:01.194507 14042 main.go:141] libmachine: Using SSH client type: native
I0915 06:30:01.194730 14042 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 06:30:01.194750 14042 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-224201 && echo "addons-224201" | sudo tee /etc/hostname
I0915 06:30:01.335848 14042 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-224201
I0915 06:30:01.335913 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:01.352266 14042 main.go:141] libmachine: Using SSH client type: native
I0915 06:30:01.352434 14042 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 06:30:01.352450 14042 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-224201' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-224201/g' /etc/hosts;
else
echo '127.0.1.1 addons-224201' | sudo tee -a /etc/hosts;
fi
fi
I0915 06:30:01.481546 14042 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0915 06:30:01.481576 14042 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19644-6048/.minikube CaCertPath:/home/jenkins/minikube-integration/19644-6048/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19644-6048/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19644-6048/.minikube}
I0915 06:30:01.481598 14042 ubuntu.go:177] setting up certificates
I0915 06:30:01.481607 14042 provision.go:84] configureAuth start
I0915 06:30:01.481648 14042 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-224201
I0915 06:30:01.497314 14042 provision.go:143] copyHostCerts
I0915 06:30:01.497395 14042 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19644-6048/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19644-6048/.minikube/ca.pem (1078 bytes)
I0915 06:30:01.497504 14042 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19644-6048/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19644-6048/.minikube/cert.pem (1123 bytes)
I0915 06:30:01.497559 14042 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19644-6048/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19644-6048/.minikube/key.pem (1679 bytes)
I0915 06:30:01.497616 14042 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19644-6048/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19644-6048/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19644-6048/.minikube/certs/ca-key.pem org=jenkins.addons-224201 san=[127.0.0.1 192.168.49.2 addons-224201 localhost minikube]
I0915 06:30:01.628613 14042 provision.go:177] copyRemoteCerts
I0915 06:30:01.628670 14042 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0915 06:30:01.628703 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:01.645337 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:01.737731 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0915 06:30:01.758106 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0915 06:30:01.777925 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0915 06:30:01.797867 14042 provision.go:87] duration metric: took 316.248547ms to configureAuth
I0915 06:30:01.797895 14042 ubuntu.go:193] setting minikube options for container-runtime
I0915 06:30:01.798073 14042 config.go:182] Loaded profile config "addons-224201": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0915 06:30:01.798127 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:01.813512 14042 main.go:141] libmachine: Using SSH client type: native
I0915 06:30:01.813699 14042 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 06:30:01.813714 14042 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0915 06:30:01.941799 14042 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0915 06:30:01.941825 14042 ubuntu.go:71] root file system type: overlay
I0915 06:30:01.941995 14042 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0915 06:30:01.942073 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:01.957749 14042 main.go:141] libmachine: Using SSH client type: native
I0915 06:30:01.957915 14042 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 06:30:01.958014 14042 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0915 06:30:02.095944 14042 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0915 06:30:02.096012 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:02.112712 14042 main.go:141] libmachine: Using SSH client type: native
I0915 06:30:02.112942 14042 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x86c560] 0x86f240 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0915 06:30:02.112967 14042 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0915 06:30:02.777007 14042 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-09-06 12:06:41.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-15 06:30:02.090838020 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0915 06:30:02.777039 14042 machine.go:96] duration metric: took 1.786152398s to provisionDockerMachine
I0915 06:30:02.777053 14042 client.go:171] duration metric: took 13.986235671s to LocalClient.Create
I0915 06:30:02.777073 14042 start.go:167] duration metric: took 13.986299247s to libmachine.API.Create "addons-224201"
I0915 06:30:02.777086 14042 start.go:293] postStartSetup for "addons-224201" (driver="docker")
I0915 06:30:02.777099 14042 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0915 06:30:02.777174 14042 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0915 06:30:02.777221 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:02.792750 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:02.886226 14042 ssh_runner.go:195] Run: cat /etc/os-release
I0915 06:30:02.889058 14042 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0915 06:30:02.889085 14042 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0915 06:30:02.889138 14042 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0915 06:30:02.889149 14042 info.go:137] Remote host: Ubuntu 22.04.4 LTS
I0915 06:30:02.889160 14042 filesync.go:126] Scanning /home/jenkins/minikube-integration/19644-6048/.minikube/addons for local assets ...
I0915 06:30:02.889217 14042 filesync.go:126] Scanning /home/jenkins/minikube-integration/19644-6048/.minikube/files for local assets ...
I0915 06:30:02.889239 14042 start.go:296] duration metric: took 112.146445ms for postStartSetup
I0915 06:30:02.889509 14042 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-224201
I0915 06:30:02.905158 14042 profile.go:143] Saving config to /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/config.json ...
I0915 06:30:02.905400 14042 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0915 06:30:02.905438 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:02.920729 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:03.010340 14042 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0915 06:30:03.014146 14042 start.go:128] duration metric: took 14.225076199s to createHost
I0915 06:30:03.014172 14042 start.go:83] releasing machines lock for "addons-224201", held for 14.225216982s
I0915 06:30:03.014223 14042 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-224201
I0915 06:30:03.029761 14042 ssh_runner.go:195] Run: cat /version.json
I0915 06:30:03.029832 14042 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0915 06:30:03.029897 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:03.029833 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:03.047007 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:03.047210 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:03.205627 14042 ssh_runner.go:195] Run: systemctl --version
I0915 06:30:03.209322 14042 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0915 06:30:03.212865 14042 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0915 06:30:03.233476 14042 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0915 06:30:03.233555 14042 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0915 06:30:03.257409 14042 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0915 06:30:03.257438 14042 start.go:495] detecting cgroup driver to use...
I0915 06:30:03.257475 14042 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0915 06:30:03.257581 14042 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0915 06:30:03.271058 14042 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0915 06:30:03.279216 14042 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0915 06:30:03.287495 14042 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0915 06:30:03.287541 14042 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0915 06:30:03.295712 14042 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0915 06:30:03.304123 14042 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0915 06:30:03.313039 14042 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0915 06:30:03.321129 14042 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0915 06:30:03.328870 14042 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0915 06:30:03.337069 14042 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0915 06:30:03.345247 14042 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0915 06:30:03.353442 14042 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0915 06:30:03.360279 14042 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0915 06:30:03.367387 14042 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 06:30:03.443145 14042 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0915 06:30:03.516095 14042 start.go:495] detecting cgroup driver to use...
I0915 06:30:03.516142 14042 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0915 06:30:03.516189 14042 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0915 06:30:03.526351 14042 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0915 06:30:03.526419 14042 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0915 06:30:03.536589 14042 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0915 06:30:03.552666 14042 ssh_runner.go:195] Run: which cri-dockerd
I0915 06:30:03.555823 14042 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0915 06:30:03.564678 14042 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0915 06:30:03.580555 14042 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0915 06:30:03.673324 14042 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0915 06:30:03.759852 14042 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0915 06:30:03.760000 14042 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0915 06:30:03.776031 14042 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 06:30:03.849089 14042 ssh_runner.go:195] Run: sudo systemctl restart docker
I0915 06:30:04.082059 14042 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0915 06:30:04.092246 14042 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0915 06:30:04.102171 14042 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0915 06:30:04.177347 14042 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0915 06:30:04.257508 14042 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 06:30:04.330014 14042 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0915 06:30:04.341878 14042 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0915 06:30:04.351216 14042 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 06:30:04.430581 14042 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0915 06:30:04.488225 14042 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0915 06:30:04.488322 14042 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0915 06:30:04.492034 14042 start.go:563] Will wait 60s for crictl version
I0915 06:30:04.492086 14042 ssh_runner.go:195] Run: which crictl
I0915 06:30:04.495137 14042 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0915 06:30:04.525837 14042 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.2.1
RuntimeApiVersion: v1
I0915 06:30:04.525915 14042 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0915 06:30:04.547536 14042 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0915 06:30:04.571606 14042 out.go:235] * Preparing Kubernetes v1.31.1 on Docker 27.2.1 ...
I0915 06:30:04.571682 14042 cli_runner.go:164] Run: docker network inspect addons-224201 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0915 06:30:04.588632 14042 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0915 06:30:04.591957 14042 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0915 06:30:04.601522 14042 kubeadm.go:883] updating cluster {Name:addons-224201 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-224201 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0915 06:30:04.601627 14042 preload.go:131] Checking if preload exists for k8s version v1.31.1 and runtime docker
I0915 06:30:04.601714 14042 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0915 06:30:04.618811 14042 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0915 06:30:04.618830 14042 docker.go:615] Images already preloaded, skipping extraction
I0915 06:30:04.618885 14042 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0915 06:30:04.636531 14042 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.1
registry.k8s.io/kube-scheduler:v1.31.1
registry.k8s.io/kube-controller-manager:v1.31.1
registry.k8s.io/kube-proxy:v1.31.1
registry.k8s.io/coredns/coredns:v1.11.3
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0915 06:30:04.636558 14042 cache_images.go:84] Images are preloaded, skipping loading
I0915 06:30:04.636567 14042 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.1 docker true true} ...
I0915 06:30:04.636688 14042 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.1/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-224201 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.1 ClusterName:addons-224201 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0915 06:30:04.636755 14042 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0915 06:30:04.680461 14042 cni.go:84] Creating CNI manager for ""
I0915 06:30:04.680486 14042 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0915 06:30:04.680499 14042 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0915 06:30:04.680522 14042 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.1 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-224201 NodeName:addons-224201 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0915 06:30:04.680663 14042 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-224201"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.1
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0915 06:30:04.680726 14042 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.1
I0915 06:30:04.688397 14042 binaries.go:44] Found k8s binaries, skipping transfer
I0915 06:30:04.688458 14042 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0915 06:30:04.696510 14042 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0915 06:30:04.712284 14042 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0915 06:30:04.727609 14042 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0915 06:30:04.742782 14042 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0915 06:30:04.745819 14042 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0915 06:30:04.755392 14042 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 06:30:04.830502 14042 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0915 06:30:04.842641 14042 certs.go:68] Setting up /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201 for IP: 192.168.49.2
I0915 06:30:04.842662 14042 certs.go:194] generating shared ca certs ...
I0915 06:30:04.842680 14042 certs.go:226] acquiring lock for ca certs: {Name:mkd8ddb0d5bbff960feae4f188ec342b3e178fe8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:04.842811 14042 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19644-6048/.minikube/ca.key
I0915 06:30:05.007571 14042 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19644-6048/.minikube/ca.crt ...
I0915 06:30:05.007601 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/ca.crt: {Name:mkc8ad9c469e878ea81b7ae92ea35e122e061a78 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:05.007760 14042 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19644-6048/.minikube/ca.key ...
I0915 06:30:05.007770 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/ca.key: {Name:mk6b8b00bbf050c20a33d75f65dfa23e12887a64 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:05.007842 14042 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19644-6048/.minikube/proxy-client-ca.key
I0915 06:30:05.122063 14042 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19644-6048/.minikube/proxy-client-ca.crt ...
I0915 06:30:05.122090 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/proxy-client-ca.crt: {Name:mk4dc1db52ebbdecb95869f2b39b7636d8402896 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:05.122257 14042 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19644-6048/.minikube/proxy-client-ca.key ...
I0915 06:30:05.122267 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/proxy-client-ca.key: {Name:mk76626adc1645dec3b3a2b44c5b1a380d049c73 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:05.122329 14042 certs.go:256] generating profile certs ...
I0915 06:30:05.122380 14042 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/client.key
I0915 06:30:05.122401 14042 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/client.crt with IP's: []
I0915 06:30:05.254955 14042 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/client.crt ...
I0915 06:30:05.254982 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/client.crt: {Name:mk11b8b16d96ee7b5f758564cb924efd52c54ac5 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:05.255132 14042 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/client.key ...
I0915 06:30:05.255141 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/client.key: {Name:mk5df4d8e1a2f951f8351f3ffb18c84007a76f51 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:05.255216 14042 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.key.e1e3eeef
I0915 06:30:05.255233 14042 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.crt.e1e3eeef with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0915 06:30:05.455734 14042 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.crt.e1e3eeef ...
I0915 06:30:05.455760 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.crt.e1e3eeef: {Name:mk3511fdab7afa1272b448c561eb0e7254515e4a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:05.455909 14042 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.key.e1e3eeef ...
I0915 06:30:05.455920 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.key.e1e3eeef: {Name:mke507e78ea79c3925decf782eb8e42b8a9ed3bd Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:05.455985 14042 certs.go:381] copying /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.crt.e1e3eeef -> /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.crt
I0915 06:30:05.456057 14042 certs.go:385] copying /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.key.e1e3eeef -> /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.key
I0915 06:30:05.456105 14042 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/proxy-client.key
I0915 06:30:05.456121 14042 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/proxy-client.crt with IP's: []
I0915 06:30:05.685951 14042 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/proxy-client.crt ...
I0915 06:30:05.685999 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/proxy-client.crt: {Name:mk9b01808dedbc59c22ef6f6d6192d56ef1240ed Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:05.686194 14042 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/proxy-client.key ...
I0915 06:30:05.686209 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/proxy-client.key: {Name:mk47277482bdad800f48ca710b89a351b082d4bd Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:05.686428 14042 certs.go:484] found cert: /home/jenkins/minikube-integration/19644-6048/.minikube/certs/ca-key.pem (1675 bytes)
I0915 06:30:05.686471 14042 certs.go:484] found cert: /home/jenkins/minikube-integration/19644-6048/.minikube/certs/ca.pem (1078 bytes)
I0915 06:30:05.686507 14042 certs.go:484] found cert: /home/jenkins/minikube-integration/19644-6048/.minikube/certs/cert.pem (1123 bytes)
I0915 06:30:05.686546 14042 certs.go:484] found cert: /home/jenkins/minikube-integration/19644-6048/.minikube/certs/key.pem (1679 bytes)
I0915 06:30:05.687133 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0915 06:30:05.708299 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1679 bytes)
I0915 06:30:05.728666 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0915 06:30:05.748369 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0915 06:30:05.768879 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0915 06:30:05.788683 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0915 06:30:05.808273 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0915 06:30:05.827971 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/profiles/addons-224201/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I0915 06:30:05.847618 14042 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19644-6048/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0915 06:30:05.867030 14042 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0915 06:30:05.881500 14042 ssh_runner.go:195] Run: openssl version
I0915 06:30:05.886095 14042 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0915 06:30:05.894086 14042 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0915 06:30:05.896843 14042 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 15 06:30 /usr/share/ca-certificates/minikubeCA.pem
I0915 06:30:05.896889 14042 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0915 06:30:05.902781 14042 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0915 06:30:05.910539 14042 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0915 06:30:05.913266 14042 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0915 06:30:05.913307 14042 kubeadm.go:392] StartCluster: {Name:addons-224201 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.45-1726358845-19644@sha256:4c67a32a16c2d4f824f00267c172fd225757ca75441e363d925dc9583137f0b0 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.1 ClusterName:addons-224201 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0915 06:30:05.913410 14042 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0915 06:30:05.929236 14042 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0915 06:30:05.937083 14042 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0915 06:30:05.944459 14042 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0915 06:30:05.944507 14042 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0915 06:30:05.951817 14042 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0915 06:30:05.951833 14042 kubeadm.go:157] found existing configuration files:
I0915 06:30:05.951875 14042 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0915 06:30:05.958719 14042 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0915 06:30:05.958765 14042 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0915 06:30:05.965456 14042 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0915 06:30:05.972170 14042 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0915 06:30:05.972214 14042 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0915 06:30:05.979038 14042 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0915 06:30:05.986069 14042 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0915 06:30:05.986117 14042 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0915 06:30:05.992688 14042 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0915 06:30:05.999773 14042 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0915 06:30:05.999829 14042 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0915 06:30:06.006683 14042 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.1:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0915 06:30:06.037845 14042 kubeadm.go:310] W0915 06:30:06.037180 1920 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0915 06:30:06.038298 14042 kubeadm.go:310] W0915 06:30:06.037770 1920 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0915 06:30:06.057125 14042 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1068-gcp\n", err: exit status 1
I0915 06:30:06.103928 14042 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0915 06:30:15.204775 14042 kubeadm.go:310] [init] Using Kubernetes version: v1.31.1
I0915 06:30:15.204843 14042 kubeadm.go:310] [preflight] Running pre-flight checks
I0915 06:30:15.204969 14042 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0915 06:30:15.205039 14042 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1068-gcp[0m
I0915 06:30:15.205077 14042 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0915 06:30:15.205152 14042 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0915 06:30:15.205195 14042 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0915 06:30:15.205236 14042 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0915 06:30:15.205290 14042 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0915 06:30:15.205364 14042 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0915 06:30:15.205418 14042 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0915 06:30:15.205456 14042 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0915 06:30:15.205530 14042 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0915 06:30:15.205597 14042 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0915 06:30:15.205667 14042 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0915 06:30:15.205756 14042 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0915 06:30:15.205840 14042 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0915 06:30:15.205893 14042 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0915 06:30:15.207346 14042 out.go:235] - Generating certificates and keys ...
I0915 06:30:15.207425 14042 kubeadm.go:310] [certs] Using existing ca certificate authority
I0915 06:30:15.207488 14042 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0915 06:30:15.207547 14042 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0915 06:30:15.207615 14042 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0915 06:30:15.207714 14042 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0915 06:30:15.207769 14042 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0915 06:30:15.207825 14042 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0915 06:30:15.207922 14042 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-224201 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0915 06:30:15.207989 14042 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0915 06:30:15.208117 14042 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-224201 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0915 06:30:15.208218 14042 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0915 06:30:15.208287 14042 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0915 06:30:15.208327 14042 kubeadm.go:310] [certs] Generating "sa" key and public key
I0915 06:30:15.208379 14042 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0915 06:30:15.208423 14042 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0915 06:30:15.208471 14042 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0915 06:30:15.208532 14042 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0915 06:30:15.208612 14042 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0915 06:30:15.208687 14042 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0915 06:30:15.208798 14042 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0915 06:30:15.208889 14042 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0915 06:30:15.210401 14042 out.go:235] - Booting up control plane ...
I0915 06:30:15.210509 14042 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0915 06:30:15.210620 14042 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0915 06:30:15.210688 14042 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0915 06:30:15.210776 14042 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0915 06:30:15.210856 14042 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0915 06:30:15.210891 14042 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0915 06:30:15.210998 14042 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0915 06:30:15.211109 14042 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0915 06:30:15.211181 14042 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 501.564711ms
I0915 06:30:15.211282 14042 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0915 06:30:15.211366 14042 kubeadm.go:310] [api-check] The API server is healthy after 4.501639353s
I0915 06:30:15.211457 14042 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0915 06:30:15.211555 14042 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0915 06:30:15.211606 14042 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0915 06:30:15.211745 14042 kubeadm.go:310] [mark-control-plane] Marking the node addons-224201 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0915 06:30:15.211789 14042 kubeadm.go:310] [bootstrap-token] Using token: np67ex.k3k8n3owif00mrbd
I0915 06:30:15.213131 14042 out.go:235] - Configuring RBAC rules ...
I0915 06:30:15.213213 14042 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0915 06:30:15.213298 14042 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0915 06:30:15.213426 14042 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0915 06:30:15.213548 14042 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0915 06:30:15.213653 14042 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0915 06:30:15.213725 14042 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0915 06:30:15.213865 14042 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0915 06:30:15.213910 14042 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0915 06:30:15.213999 14042 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0915 06:30:15.214010 14042 kubeadm.go:310]
I0915 06:30:15.214078 14042 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0915 06:30:15.214085 14042 kubeadm.go:310]
I0915 06:30:15.214159 14042 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0915 06:30:15.214170 14042 kubeadm.go:310]
I0915 06:30:15.214204 14042 kubeadm.go:310] mkdir -p $HOME/.kube
I0915 06:30:15.214290 14042 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0915 06:30:15.214341 14042 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0915 06:30:15.214348 14042 kubeadm.go:310]
I0915 06:30:15.214396 14042 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0915 06:30:15.214402 14042 kubeadm.go:310]
I0915 06:30:15.214440 14042 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0915 06:30:15.214446 14042 kubeadm.go:310]
I0915 06:30:15.214495 14042 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0915 06:30:15.214562 14042 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0915 06:30:15.214623 14042 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0915 06:30:15.214629 14042 kubeadm.go:310]
I0915 06:30:15.214699 14042 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0915 06:30:15.214767 14042 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0915 06:30:15.214776 14042 kubeadm.go:310]
I0915 06:30:15.214856 14042 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token np67ex.k3k8n3owif00mrbd \
I0915 06:30:15.214978 14042 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:ace8bb1ca27835134178bb8e2bb49177450bf635c1c2b41ba4aa5b89ed07cdb6 \
I0915 06:30:15.215011 14042 kubeadm.go:310] --control-plane
I0915 06:30:15.215020 14042 kubeadm.go:310]
I0915 06:30:15.215099 14042 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0915 06:30:15.215107 14042 kubeadm.go:310]
I0915 06:30:15.215169 14042 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token np67ex.k3k8n3owif00mrbd \
I0915 06:30:15.215272 14042 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:ace8bb1ca27835134178bb8e2bb49177450bf635c1c2b41ba4aa5b89ed07cdb6
I0915 06:30:15.215287 14042 cni.go:84] Creating CNI manager for ""
I0915 06:30:15.215303 14042 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0915 06:30:15.216763 14042 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0915 06:30:15.218047 14042 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0915 06:30:15.225843 14042 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0915 06:30:15.240720 14042 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0915 06:30:15.240806 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:15.240828 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-224201 minikube.k8s.io/updated_at=2024_09_15T06_30_15_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=7a3ca67a20528f5dabbb456e8e4ce542b58ef23a minikube.k8s.io/name=addons-224201 minikube.k8s.io/primary=true
I0915 06:30:15.247276 14042 ops.go:34] apiserver oom_adj: -16
I0915 06:30:15.322072 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:15.822524 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:16.322189 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:16.822114 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:17.322685 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:17.822115 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:18.322811 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:18.823174 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:19.322732 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:19.823148 14042 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.1/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0915 06:30:19.885333 14042 kubeadm.go:1113] duration metric: took 4.644583129s to wait for elevateKubeSystemPrivileges
I0915 06:30:19.885377 14042 kubeadm.go:394] duration metric: took 13.972071624s to StartCluster
I0915 06:30:19.885401 14042 settings.go:142] acquiring lock: {Name:mk489e9f2fdb278fe067a4c5b555b1a7e1d2c054 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:19.885521 14042 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19644-6048/kubeconfig
I0915 06:30:19.886072 14042 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19644-6048/kubeconfig: {Name:mkb5555886ac5bc47bf5e2da5f3b9591e7fdc0a9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0915 06:30:19.886354 14042 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0915 06:30:19.886381 14042 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.1 ContainerRuntime:docker ControlPlane:true Worker:true}
I0915 06:30:19.886425 14042 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0915 06:30:19.886552 14042 addons.go:69] Setting cloud-spanner=true in profile "addons-224201"
I0915 06:30:19.886565 14042 addons.go:69] Setting yakd=true in profile "addons-224201"
I0915 06:30:19.886578 14042 addons.go:234] Setting addon cloud-spanner=true in "addons-224201"
I0915 06:30:19.886579 14042 addons.go:234] Setting addon yakd=true in "addons-224201"
I0915 06:30:19.886582 14042 addons.go:69] Setting helm-tiller=true in profile "addons-224201"
I0915 06:30:19.886607 14042 addons.go:234] Setting addon helm-tiller=true in "addons-224201"
I0915 06:30:19.886608 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.886614 14042 config.go:182] Loaded profile config "addons-224201": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0915 06:30:19.886625 14042 addons.go:69] Setting ingress-dns=true in profile "addons-224201"
I0915 06:30:19.886635 14042 addons.go:234] Setting addon ingress-dns=true in "addons-224201"
I0915 06:30:19.886645 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.886616 14042 addons.go:69] Setting ingress=true in profile "addons-224201"
I0915 06:30:19.886675 14042 addons.go:234] Setting addon ingress=true in "addons-224201"
I0915 06:30:19.886686 14042 addons.go:69] Setting inspektor-gadget=true in profile "addons-224201"
I0915 06:30:19.886699 14042 addons.go:234] Setting addon inspektor-gadget=true in "addons-224201"
I0915 06:30:19.886707 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.886716 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.886752 14042 addons.go:69] Setting default-storageclass=true in profile "addons-224201"
I0915 06:30:19.886782 14042 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-224201"
I0915 06:30:19.887080 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.887142 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.887156 14042 addons.go:69] Setting metrics-server=true in profile "addons-224201"
I0915 06:30:19.887161 14042 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-224201"
I0915 06:30:19.887174 14042 addons.go:234] Setting addon metrics-server=true in "addons-224201"
I0915 06:30:19.887196 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.887209 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.887220 14042 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-224201"
I0915 06:30:19.887232 14042 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-224201"
I0915 06:30:19.887257 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.887147 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.887613 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.887673 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.886678 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.888242 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.887147 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.889105 14042 out.go:177] * Verifying Kubernetes components...
I0915 06:30:19.887197 14042 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-224201"
I0915 06:30:19.887209 14042 addons.go:69] Setting storage-provisioner=true in profile "addons-224201"
I0915 06:30:19.889294 14042 addons.go:69] Setting gcp-auth=true in profile "addons-224201"
I0915 06:30:19.889330 14042 addons.go:69] Setting registry=true in profile "addons-224201"
I0915 06:30:19.886608 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.889365 14042 addons.go:69] Setting volumesnapshots=true in profile "addons-224201"
I0915 06:30:19.889373 14042 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-224201"
I0915 06:30:19.889382 14042 addons.go:69] Setting volcano=true in profile "addons-224201"
I0915 06:30:19.892525 14042 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0915 06:30:19.892703 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.893566 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.894112 14042 addons.go:234] Setting addon registry=true in "addons-224201"
I0915 06:30:19.894151 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.894596 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.894936 14042 mustload.go:65] Loading cluster: addons-224201
I0915 06:30:19.895142 14042 config.go:182] Loaded profile config "addons-224201": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.1
I0915 06:30:19.895384 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.896978 14042 addons.go:234] Setting addon storage-provisioner=true in "addons-224201"
I0915 06:30:19.897158 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.897211 14042 addons.go:234] Setting addon volcano=true in "addons-224201"
I0915 06:30:19.897578 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.897834 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.897221 14042 addons.go:234] Setting addon volumesnapshots=true in "addons-224201"
I0915 06:30:19.899210 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.897234 14042 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-224201"
I0915 06:30:19.905369 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.919479 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.919873 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.921718 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.929140 14042 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0915 06:30:19.931011 14042 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0915 06:30:19.931125 14042 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0915 06:30:19.931142 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0915 06:30:19.931192 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:19.932503 14042 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0915 06:30:19.932526 14042 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0915 06:30:19.932586 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:19.934408 14042 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0915 06:30:19.934522 14042 out.go:177] - Using image docker.io/registry:2.8.3
I0915 06:30:19.935704 14042 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0915 06:30:19.935725 14042 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0915 06:30:19.935783 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:19.936913 14042 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0915 06:30:19.938705 14042 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0915 06:30:19.938725 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0915 06:30:19.938772 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:19.944345 14042 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0915 06:30:19.945518 14042 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0915 06:30:19.945682 14042 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0915 06:30:19.945698 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0915 06:30:19.945745 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:19.947952 14042 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0915 06:30:19.950164 14042 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0915 06:30:19.951633 14042 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0915 06:30:19.951652 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0915 06:30:19.951703 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:19.974261 14042 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0915 06:30:19.975535 14042 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0915 06:30:19.975563 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0915 06:30:19.975618 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:19.980969 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:19.982588 14042 addons.go:234] Setting addon default-storageclass=true in "addons-224201"
I0915 06:30:19.982636 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.983113 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.983907 14042 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0915 06:30:19.985490 14042 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0915 06:30:19.985641 14042 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0915 06:30:19.985654 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0915 06:30:19.985700 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:19.985939 14042 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0915 06:30:19.989585 14042 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-224201"
I0915 06:30:19.998231 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:19.998703 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:19.994625 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:20.000032 14042 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0915 06:30:20.000056 14042 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0915 06:30:20.000102 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:20.002046 14042 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0915 06:30:20.003298 14042 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0915 06:30:20.003314 14042 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0915 06:30:20.003369 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:20.004011 14042 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0915 06:30:20.004066 14042 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0915 06:30:20.005583 14042 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0915 06:30:20.005597 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0915 06:30:20.005637 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:20.005790 14042 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0915 06:30:20.006956 14042 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0915 06:30:20.008131 14042 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0915 06:30:20.008946 14042 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0915 06:30:20.010655 14042 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0915 06:30:20.010676 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0915 06:30:20.010721 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:20.017092 14042 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0915 06:30:20.018488 14042 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0915 06:30:20.019735 14042 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0915 06:30:20.020727 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.021933 14042 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0915 06:30:20.024893 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.026034 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.028574 14042 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0915 06:30:20.029828 14042 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0915 06:30:20.029845 14042 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0915 06:30:20.029900 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:20.038416 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.043750 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.044374 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.046041 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.049113 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.049539 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.050746 14042 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0915 06:30:20.050763 14042 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0915 06:30:20.050805 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:20.052844 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.052973 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.058714 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.060089 14042 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0915 06:30:20.065503 14042 out.go:177] - Using image docker.io/busybox:stable
I0915 06:30:20.067142 14042 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0915 06:30:20.067163 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0915 06:30:20.067212 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
W0915 06:30:20.069324 14042 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0915 06:30:20.069351 14042 retry.go:31] will retry after 135.696348ms: ssh: handshake failed: EOF
I0915 06:30:20.089608 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:20.097532 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
W0915 06:30:20.158473 14042 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0915 06:30:20.158503 14042 retry.go:31] will retry after 311.839574ms: ssh: handshake failed: EOF
I0915 06:30:20.260084 14042 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0915 06:30:20.260210 14042 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0915 06:30:20.378504 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0915 06:30:20.388612 14042 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0915 06:30:20.388672 14042 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0915 06:30:20.471527 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0915 06:30:20.476190 14042 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0915 06:30:20.476213 14042 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0915 06:30:20.577092 14042 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0915 06:30:20.577187 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0915 06:30:20.578516 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0915 06:30:20.579857 14042 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0915 06:30:20.579872 14042 ssh_runner.go:362] scp helm-tiller/helm-tiller-rbac.yaml --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0915 06:30:20.581118 14042 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0915 06:30:20.581135 14042 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0915 06:30:20.585139 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0915 06:30:20.660795 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0915 06:30:20.667008 14042 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0915 06:30:20.667078 14042 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0915 06:30:20.669071 14042 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0915 06:30:20.669122 14042 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0915 06:30:20.759056 14042 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0915 06:30:20.759136 14042 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0915 06:30:20.769777 14042 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0915 06:30:20.769861 14042 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0915 06:30:20.771513 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0915 06:30:20.773842 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0915 06:30:20.777935 14042 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0915 06:30:20.777958 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0915 06:30:20.780652 14042 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0915 06:30:20.780689 14042 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0915 06:30:20.959346 14042 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0915 06:30:20.959430 14042 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0915 06:30:20.960238 14042 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0915 06:30:20.960284 14042 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0915 06:30:21.060244 14042 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0915 06:30:21.060290 14042 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0915 06:30:21.063879 14042 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0915 06:30:21.063906 14042 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0915 06:30:21.261582 14042 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0915 06:30:21.261680 14042 ssh_runner.go:362] scp helm-tiller/helm-tiller-svc.yaml --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0915 06:30:21.263534 14042 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0915 06:30:21.263616 14042 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0915 06:30:21.272307 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0915 06:30:21.280338 14042 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0915 06:30:21.280370 14042 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0915 06:30:21.372919 14042 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0915 06:30:21.372946 14042 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0915 06:30:21.460729 14042 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0915 06:30:21.460832 14042 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0915 06:30:21.468490 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0915 06:30:21.561155 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0915 06:30:21.659696 14042 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0915 06:30:21.659797 14042 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0915 06:30:21.864821 14042 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0915 06:30:21.864914 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0915 06:30:21.958809 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0915 06:30:21.960523 14042 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0915 06:30:21.960586 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0915 06:30:21.964381 14042 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.1/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.704141192s)
I0915 06:30:21.964493 14042 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0915 06:30:21.964461 14042 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.704350315s)
I0915 06:30:21.965761 14042 node_ready.go:35] waiting up to 6m0s for node "addons-224201" to be "Ready" ...
I0915 06:30:21.969404 14042 node_ready.go:49] node "addons-224201" has status "Ready":"True"
I0915 06:30:21.969490 14042 node_ready.go:38] duration metric: took 3.663789ms for node "addons-224201" to be "Ready" ...
I0915 06:30:21.969517 14042 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0915 06:30:21.978623 14042 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace to be "Ready" ...
I0915 06:30:22.377735 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0915 06:30:22.468150 14042 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-224201" context rescaled to 1 replicas
I0915 06:30:22.468687 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0915 06:30:22.476290 14042 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0915 06:30:22.476352 14042 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0915 06:30:22.664703 14042 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0915 06:30:22.664751 14042 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0915 06:30:22.872284 14042 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0915 06:30:22.872325 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0915 06:30:22.976285 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (2.597687534s)
I0915 06:30:23.261418 14042 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0915 06:30:23.261524 14042 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0915 06:30:23.478311 14042 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0915 06:30:23.478398 14042 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0915 06:30:23.762467 14042 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0915 06:30:23.762496 14042 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0915 06:30:24.060580 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:24.265006 14042 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0915 06:30:24.265050 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0915 06:30:24.565498 14042 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0915 06:30:24.565596 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0915 06:30:24.872441 14042 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0915 06:30:24.872518 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0915 06:30:25.278758 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0915 06:30:25.868196 14042 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0915 06:30:25.868275 14042 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0915 06:30:26.077049 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:26.562104 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0915 06:30:27.160445 14042 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0915 06:30:27.160607 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:27.190904 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:28.270375 14042 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0915 06:30:28.569392 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:28.672877 14042 addons.go:234] Setting addon gcp-auth=true in "addons-224201"
I0915 06:30:28.672938 14042 host.go:66] Checking if "addons-224201" exists ...
I0915 06:30:28.673449 14042 cli_runner.go:164] Run: docker container inspect addons-224201 --format={{.State.Status}}
I0915 06:30:28.695751 14042 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0915 06:30:28.695795 14042 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-224201
I0915 06:30:28.710728 14042 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19644-6048/.minikube/machines/addons-224201/id_rsa Username:docker}
I0915 06:30:29.462613 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (8.991039277s)
I0915 06:30:29.462683 14042 addons.go:475] Verifying addon ingress=true in "addons-224201"
I0915 06:30:29.469315 14042 out.go:177] * Verifying ingress addon...
I0915 06:30:29.471893 14042 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0915 06:30:29.478084 14042 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0915 06:30:29.478108 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:29.977064 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:30.476112 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:30.978351 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:31.061232 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:31.476828 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:31.977299 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:32.473766 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.895209917s)
I0915 06:30:32.473836 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (11.888676555s)
I0915 06:30:32.474078 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (11.813194483s)
I0915 06:30:32.474131 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (11.70254543s)
I0915 06:30:32.474192 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (11.700317505s)
I0915 06:30:32.474256 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (11.201926588s)
I0915 06:30:32.474443 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (11.005868174s)
I0915 06:30:32.474456 14042 addons.go:475] Verifying addon metrics-server=true in "addons-224201"
I0915 06:30:32.474504 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (10.913256735s)
I0915 06:30:32.474514 14042 addons.go:475] Verifying addon registry=true in "addons-224201"
I0915 06:30:32.474873 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (10.515959477s)
I0915 06:30:32.474990 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (10.097138643s)
I0915 06:30:32.475120 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (10.006379018s)
W0915 06:30:32.475170 14042 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0915 06:30:32.475201 14042 retry.go:31] will retry after 336.443743ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0915 06:30:32.475305 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (7.196433468s)
I0915 06:30:32.476293 14042 out.go:177] * Verifying registry addon...
I0915 06:30:32.476292 14042 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-224201 service yakd-dashboard -n yakd-dashboard
I0915 06:30:32.478614 14042 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0915 06:30:32.478862 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:32.481961 14042 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0915 06:30:32.482028 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
W0915 06:30:32.565207 14042 out.go:270] ! Enabling 'storage-provisioner-rancher' returned an error: running callbacks: [Error making local-path the default storage class: Error while marking storage class local-path as default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
I0915 06:30:32.812662 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0915 06:30:33.080028 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:33.080303 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:33.159074 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:33.477103 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:33.483945 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:33.662331 14042 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (4.966547918s)
I0915 06:30:33.662467 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (7.100207492s)
I0915 06:30:33.662495 14042 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-224201"
I0915 06:30:33.663997 14042 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0915 06:30:33.664053 14042 out.go:177] * Verifying csi-hostpath-driver addon...
I0915 06:30:33.665835 14042 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0915 06:30:33.666673 14042 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0915 06:30:33.667067 14042 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0915 06:30:33.667115 14042 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0915 06:30:33.670817 14042 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0915 06:30:33.670837 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:33.759235 14042 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0915 06:30:33.759263 14042 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0915 06:30:33.780395 14042 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0915 06:30:33.780417 14042 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0915 06:30:33.858643 14042 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0915 06:30:33.975935 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:33.982912 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:34.172249 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:34.476340 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:34.482355 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:34.672053 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:34.976942 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:34.982146 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.16944366s)
I0915 06:30:34.983913 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:35.172314 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:35.259477 14042 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.1/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.400794422s)
I0915 06:30:35.260988 14042 addons.go:475] Verifying addon gcp-auth=true in "addons-224201"
I0915 06:30:35.262691 14042 out.go:177] * Verifying gcp-auth addon...
I0915 06:30:35.265382 14042 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0915 06:30:35.270766 14042 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0915 06:30:35.477241 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:35.482045 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:35.483883 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:35.670966 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:35.978522 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:35.981469 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:36.171323 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:36.476664 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:36.482579 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:36.671597 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:36.976191 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:36.982415 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:37.171021 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:37.476391 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:37.482283 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:37.483953 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:37.670884 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:37.976041 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:37.981925 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:38.171247 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:38.596516 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:38.597069 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:38.670439 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:38.976279 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:38.982467 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:39.170633 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:39.476264 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:39.482251 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:39.484029 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:39.671758 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:39.976648 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:39.981550 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:40.170628 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:40.476236 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:40.482023 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:40.671565 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:40.976647 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:40.981452 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:41.171010 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:41.475937 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:41.481686 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:41.671894 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:41.976155 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:41.981680 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:41.983510 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:42.171319 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:42.475567 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:42.481336 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:42.671051 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:42.976084 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:42.981903 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:43.172108 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:43.475930 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:43.481424 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:43.672250 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:43.975986 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:43.981730 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:44.171786 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:44.475659 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:44.481613 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:44.483589 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:44.670957 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:44.975875 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:44.981809 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:45.171935 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:45.476748 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:45.481500 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:45.671598 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:45.976395 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:45.982514 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:46.171218 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:46.544868 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:46.545161 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:46.546454 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:46.671475 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:46.975804 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:46.981683 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:47.171277 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:47.477271 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:47.481614 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:47.671355 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:47.976048 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:47.981791 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:48.172361 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:48.477011 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:48.481801 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:48.671860 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:48.976342 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:48.981868 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:48.983662 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:49.172109 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:49.475746 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:49.484352 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:49.671131 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:49.976812 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:49.981862 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:50.170483 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:50.477007 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:50.482422 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:50.671439 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:50.976519 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:50.981536 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:51.171789 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:51.475976 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:51.481912 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:51.484035 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:51.670972 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:51.975728 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:51.981415 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:52.171380 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:52.475850 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:52.481744 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:52.670566 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:52.975765 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:52.981651 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:53.171285 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:53.476367 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:53.481910 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:53.484296 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:53.671170 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:53.975917 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:53.981782 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0915 06:30:54.170805 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:54.476809 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:54.481919 14042 kapi.go:107] duration metric: took 22.003300264s to wait for kubernetes.io/minikube-addons=registry ...
I0915 06:30:54.671697 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:54.976581 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:55.171416 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:55.476844 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:55.671480 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:55.977196 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:55.984100 14042 pod_ready.go:103] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"False"
I0915 06:30:56.170923 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:56.476410 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:56.671353 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:57.065603 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:57.171114 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:57.476824 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:57.671801 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:57.975958 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:57.983882 14042 pod_ready.go:93] pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace has status "Ready":"True"
I0915 06:30:57.983911 14042 pod_ready.go:82] duration metric: took 36.005218653s for pod "coredns-7c65d6cfc9-q4qbc" in "kube-system" namespace to be "Ready" ...
I0915 06:30:57.983923 14042 pod_ready.go:79] waiting up to 6m0s for pod "coredns-7c65d6cfc9-rr92z" in "kube-system" namespace to be "Ready" ...
I0915 06:30:57.985484 14042 pod_ready.go:98] error getting pod "coredns-7c65d6cfc9-rr92z" in "kube-system" namespace (skipping!): pods "coredns-7c65d6cfc9-rr92z" not found
I0915 06:30:57.985508 14042 pod_ready.go:82] duration metric: took 1.576596ms for pod "coredns-7c65d6cfc9-rr92z" in "kube-system" namespace to be "Ready" ...
E0915 06:30:57.985520 14042 pod_ready.go:67] WaitExtra: waitPodCondition: error getting pod "coredns-7c65d6cfc9-rr92z" in "kube-system" namespace (skipping!): pods "coredns-7c65d6cfc9-rr92z" not found
I0915 06:30:57.985529 14042 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-224201" in "kube-system" namespace to be "Ready" ...
I0915 06:30:57.989251 14042 pod_ready.go:93] pod "etcd-addons-224201" in "kube-system" namespace has status "Ready":"True"
I0915 06:30:57.989268 14042 pod_ready.go:82] duration metric: took 3.729491ms for pod "etcd-addons-224201" in "kube-system" namespace to be "Ready" ...
I0915 06:30:57.989277 14042 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-224201" in "kube-system" namespace to be "Ready" ...
I0915 06:30:57.993258 14042 pod_ready.go:93] pod "kube-apiserver-addons-224201" in "kube-system" namespace has status "Ready":"True"
I0915 06:30:57.993277 14042 pod_ready.go:82] duration metric: took 3.992053ms for pod "kube-apiserver-addons-224201" in "kube-system" namespace to be "Ready" ...
I0915 06:30:57.993287 14042 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-224201" in "kube-system" namespace to be "Ready" ...
I0915 06:30:57.997137 14042 pod_ready.go:93] pod "kube-controller-manager-addons-224201" in "kube-system" namespace has status "Ready":"True"
I0915 06:30:57.997156 14042 pod_ready.go:82] duration metric: took 3.857706ms for pod "kube-controller-manager-addons-224201" in "kube-system" namespace to be "Ready" ...
I0915 06:30:57.997166 14042 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-g8mtm" in "kube-system" namespace to be "Ready" ...
I0915 06:30:58.171043 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:58.182535 14042 pod_ready.go:93] pod "kube-proxy-g8mtm" in "kube-system" namespace has status "Ready":"True"
I0915 06:30:58.182554 14042 pod_ready.go:82] duration metric: took 185.382064ms for pod "kube-proxy-g8mtm" in "kube-system" namespace to be "Ready" ...
I0915 06:30:58.182569 14042 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-224201" in "kube-system" namespace to be "Ready" ...
I0915 06:30:58.476360 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:58.582253 14042 pod_ready.go:93] pod "kube-scheduler-addons-224201" in "kube-system" namespace has status "Ready":"True"
I0915 06:30:58.582280 14042 pod_ready.go:82] duration metric: took 399.701979ms for pod "kube-scheduler-addons-224201" in "kube-system" namespace to be "Ready" ...
I0915 06:30:58.582291 14042 pod_ready.go:79] waiting up to 6m0s for pod "nvidia-device-plugin-daemonset-5qsk2" in "kube-system" namespace to be "Ready" ...
I0915 06:30:58.671204 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:58.976106 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:58.982529 14042 pod_ready.go:93] pod "nvidia-device-plugin-daemonset-5qsk2" in "kube-system" namespace has status "Ready":"True"
I0915 06:30:58.982552 14042 pod_ready.go:82] duration metric: took 400.251714ms for pod "nvidia-device-plugin-daemonset-5qsk2" in "kube-system" namespace to be "Ready" ...
I0915 06:30:58.982566 14042 pod_ready.go:39] duration metric: took 37.013006195s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0915 06:30:58.982589 14042 api_server.go:52] waiting for apiserver process to appear ...
I0915 06:30:58.982651 14042 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0915 06:30:58.996521 14042 api_server.go:72] duration metric: took 39.110101301s to wait for apiserver process to appear ...
I0915 06:30:58.996543 14042 api_server.go:88] waiting for apiserver healthz status ...
I0915 06:30:58.996570 14042 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0915 06:30:59.001441 14042 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0915 06:30:59.002382 14042 api_server.go:141] control plane version: v1.31.1
I0915 06:30:59.002408 14042 api_server.go:131] duration metric: took 5.856494ms to wait for apiserver health ...
I0915 06:30:59.002417 14042 system_pods.go:43] waiting for kube-system pods to appear ...
I0915 06:30:59.171586 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:59.188068 14042 system_pods.go:59] 18 kube-system pods found
I0915 06:30:59.188116 14042 system_pods.go:61] "coredns-7c65d6cfc9-q4qbc" [97df761a-6a9c-4ff2-be51-2de24a03619f] Running
I0915 06:30:59.188132 14042 system_pods.go:61] "csi-hostpath-attacher-0" [3e3400eb-30fd-4e0b-a4ea-bc1239886d4c] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0915 06:30:59.188144 14042 system_pods.go:61] "csi-hostpath-resizer-0" [ccdbe0fe-a57d-4766-9761-84d38acb3478] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0915 06:30:59.188159 14042 system_pods.go:61] "csi-hostpathplugin-hbf26" [612d1431-1630-44b5-8992-864ad9bdecc5] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0915 06:30:59.188169 14042 system_pods.go:61] "etcd-addons-224201" [ef963b52-738f-406c-b131-eaf94a440d30] Running
I0915 06:30:59.188176 14042 system_pods.go:61] "kube-apiserver-addons-224201" [f1c8de9b-eda9-4cb0-974f-b4ba86b6477f] Running
I0915 06:30:59.188187 14042 system_pods.go:61] "kube-controller-manager-addons-224201" [a6420131-8474-4116-8aee-7f5fccc03d4f] Running
I0915 06:30:59.188196 14042 system_pods.go:61] "kube-ingress-dns-minikube" [d8b15c1c-3344-44fb-ba6f-0b6bd6d044eb] Running
I0915 06:30:59.188202 14042 system_pods.go:61] "kube-proxy-g8mtm" [828188ad-a787-4814-bfc0-0000f0267eee] Running
I0915 06:30:59.188210 14042 system_pods.go:61] "kube-scheduler-addons-224201" [cc23293e-893e-402c-8188-3f088a4ed05c] Running
I0915 06:30:59.188216 14042 system_pods.go:61] "metrics-server-84c5f94fbc-5m4rs" [ce395e67-6b8d-4ebe-b259-b73c0e1c6bde] Running
I0915 06:30:59.188220 14042 system_pods.go:61] "nvidia-device-plugin-daemonset-5qsk2" [8938be57-4e5d-4e04-9565-f31703be09d8] Running
I0915 06:30:59.188226 14042 system_pods.go:61] "registry-66c9cd494c-t9hlq" [48d3ee3d-2a77-4bbb-9492-d6cb8accc5e0] Running
I0915 06:30:59.188229 14042 system_pods.go:61] "registry-proxy-k8hwr" [8343911e-7ef0-4555-a6d5-acefa24b2ae3] Running
I0915 06:30:59.188235 14042 system_pods.go:61] "snapshot-controller-56fcc65765-69w8z" [2b3ff722-5ed4-4d83-a9d1-d03a7afa34d1] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0915 06:30:59.188241 14042 system_pods.go:61] "snapshot-controller-56fcc65765-tmnpw" [cfa9ed34-cd29-4f33-a973-b18041c2e794] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0915 06:30:59.188245 14042 system_pods.go:61] "storage-provisioner" [a4519631-975a-4e32-8999-1da10d156060] Running
I0915 06:30:59.188249 14042 system_pods.go:61] "tiller-deploy-b48cc5f79-zjccb" [2764ae41-ccca-47cf-ae9e-9e3caf60ab81] Running
I0915 06:30:59.188255 14042 system_pods.go:74] duration metric: took 185.832528ms to wait for pod list to return data ...
I0915 06:30:59.188265 14042 default_sa.go:34] waiting for default service account to be created ...
I0915 06:30:59.382296 14042 default_sa.go:45] found service account: "default"
I0915 06:30:59.382321 14042 default_sa.go:55] duration metric: took 194.049318ms for default service account to be created ...
I0915 06:30:59.382335 14042 system_pods.go:116] waiting for k8s-apps to be running ...
I0915 06:30:59.574588 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:30:59.588384 14042 system_pods.go:86] 18 kube-system pods found
I0915 06:30:59.588417 14042 system_pods.go:89] "coredns-7c65d6cfc9-q4qbc" [97df761a-6a9c-4ff2-be51-2de24a03619f] Running
I0915 06:30:59.588431 14042 system_pods.go:89] "csi-hostpath-attacher-0" [3e3400eb-30fd-4e0b-a4ea-bc1239886d4c] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0915 06:30:59.588440 14042 system_pods.go:89] "csi-hostpath-resizer-0" [ccdbe0fe-a57d-4766-9761-84d38acb3478] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0915 06:30:59.588454 14042 system_pods.go:89] "csi-hostpathplugin-hbf26" [612d1431-1630-44b5-8992-864ad9bdecc5] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0915 06:30:59.588464 14042 system_pods.go:89] "etcd-addons-224201" [ef963b52-738f-406c-b131-eaf94a440d30] Running
I0915 06:30:59.588473 14042 system_pods.go:89] "kube-apiserver-addons-224201" [f1c8de9b-eda9-4cb0-974f-b4ba86b6477f] Running
I0915 06:30:59.588482 14042 system_pods.go:89] "kube-controller-manager-addons-224201" [a6420131-8474-4116-8aee-7f5fccc03d4f] Running
I0915 06:30:59.588492 14042 system_pods.go:89] "kube-ingress-dns-minikube" [d8b15c1c-3344-44fb-ba6f-0b6bd6d044eb] Running
I0915 06:30:59.588501 14042 system_pods.go:89] "kube-proxy-g8mtm" [828188ad-a787-4814-bfc0-0000f0267eee] Running
I0915 06:30:59.588507 14042 system_pods.go:89] "kube-scheduler-addons-224201" [cc23293e-893e-402c-8188-3f088a4ed05c] Running
I0915 06:30:59.588516 14042 system_pods.go:89] "metrics-server-84c5f94fbc-5m4rs" [ce395e67-6b8d-4ebe-b259-b73c0e1c6bde] Running
I0915 06:30:59.588525 14042 system_pods.go:89] "nvidia-device-plugin-daemonset-5qsk2" [8938be57-4e5d-4e04-9565-f31703be09d8] Running
I0915 06:30:59.588534 14042 system_pods.go:89] "registry-66c9cd494c-t9hlq" [48d3ee3d-2a77-4bbb-9492-d6cb8accc5e0] Running
I0915 06:30:59.588542 14042 system_pods.go:89] "registry-proxy-k8hwr" [8343911e-7ef0-4555-a6d5-acefa24b2ae3] Running
I0915 06:30:59.588551 14042 system_pods.go:89] "snapshot-controller-56fcc65765-69w8z" [2b3ff722-5ed4-4d83-a9d1-d03a7afa34d1] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0915 06:30:59.588561 14042 system_pods.go:89] "snapshot-controller-56fcc65765-tmnpw" [cfa9ed34-cd29-4f33-a973-b18041c2e794] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0915 06:30:59.588568 14042 system_pods.go:89] "storage-provisioner" [a4519631-975a-4e32-8999-1da10d156060] Running
I0915 06:30:59.588573 14042 system_pods.go:89] "tiller-deploy-b48cc5f79-zjccb" [2764ae41-ccca-47cf-ae9e-9e3caf60ab81] Running
I0915 06:30:59.588581 14042 system_pods.go:126] duration metric: took 206.241155ms to wait for k8s-apps to be running ...
I0915 06:30:59.588591 14042 system_svc.go:44] waiting for kubelet service to be running ....
I0915 06:30:59.588630 14042 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0915 06:30:59.599272 14042 system_svc.go:56] duration metric: took 10.66068ms WaitForService to wait for kubelet
I0915 06:30:59.599313 14042 kubeadm.go:582] duration metric: took 39.712897347s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0915 06:30:59.599334 14042 node_conditions.go:102] verifying NodePressure condition ...
I0915 06:30:59.670965 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:30:59.782198 14042 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0915 06:30:59.782223 14042 node_conditions.go:123] node cpu capacity is 8
I0915 06:30:59.782235 14042 node_conditions.go:105] duration metric: took 182.896816ms to run NodePressure ...
I0915 06:30:59.782246 14042 start.go:241] waiting for startup goroutines ...
I0915 06:30:59.976298 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:00.172079 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:00.476541 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:00.671911 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:00.976104 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:01.170963 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:01.476645 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:01.672514 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:01.976562 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:02.171083 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:02.476572 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:02.670799 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:02.976582 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:03.171424 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:03.476857 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:03.671036 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:03.975718 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:04.171163 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:04.476722 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:04.670466 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:04.976043 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:05.171598 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:05.475888 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:05.670635 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:05.975715 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:06.171039 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:06.476218 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:06.671484 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:06.976114 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:07.171256 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:07.476772 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:07.671418 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:07.978160 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:08.170980 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:08.476270 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:08.671290 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:08.975970 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:09.170680 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:09.476098 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:09.671188 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:09.976625 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:10.171538 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:10.477121 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:10.671032 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:10.976401 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:11.171949 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:11.475890 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:11.671416 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:11.975513 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:12.171333 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:12.476442 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:12.672035 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:12.976372 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:13.171577 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:13.476331 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:13.673866 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:13.986118 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:14.172490 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:14.476624 14042 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0915 06:31:14.672093 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:14.981832 14042 kapi.go:107] duration metric: took 45.509939056s to wait for app.kubernetes.io/name=ingress-nginx ...
I0915 06:31:15.174579 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:15.671599 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:16.171853 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:16.671251 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:17.171362 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:17.671113 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:18.171332 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:18.671284 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:19.171132 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:19.670741 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0915 06:31:20.170475 14042 kapi.go:107] duration metric: took 46.503803508s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0915 06:31:57.768443 14042 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0915 06:31:57.768462 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:31:58.268402 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:31:58.768376 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:31:59.268298 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:31:59.767885 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:00.268704 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:00.768632 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:01.268537 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:01.768267 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:02.267761 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:02.768842 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:03.268673 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:03.768790 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:04.268439 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:04.769346 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:05.268180 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:05.768927 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:06.268805 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:06.769203 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:07.269295 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:07.768279 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:08.269596 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:08.768567 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:09.268425 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:09.768467 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:10.268671 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:10.768707 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:11.268312 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:11.767793 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:12.268502 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:12.768468 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:13.268544 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:13.768456 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:14.268076 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:14.769004 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:15.269213 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:15.768297 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:16.268086 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:16.768840 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:17.268900 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:17.768633 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:18.268769 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:18.768798 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:19.268508 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:19.767941 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:20.269039 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:20.769022 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:21.268696 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:21.767938 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:22.268837 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:22.768943 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:23.269159 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:23.769186 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:24.269416 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:24.768446 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:25.268534 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:25.768253 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:26.267831 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:26.768809 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:27.268844 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:27.768666 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:28.268964 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:28.768990 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:29.268975 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:29.768685 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:30.268938 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:30.769185 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:31.268780 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:31.768332 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:32.268163 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:32.769299 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:33.268459 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:33.768631 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:34.268554 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:34.769614 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:35.268352 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:35.768347 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:36.268005 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:36.769340 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:37.268731 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:37.768925 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:38.268991 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:38.769332 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:39.268501 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:39.768240 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:40.269424 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:40.768203 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:41.269068 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:41.768970 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:42.268752 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:42.768729 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:43.268556 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:43.768277 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:44.267948 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:44.769104 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:45.268807 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:45.768868 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:46.268687 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:46.769473 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:47.268490 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:47.769078 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:48.269255 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:48.768273 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:49.268833 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:49.768024 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:50.268763 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:50.768558 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:51.267669 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:51.768364 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:52.267880 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:52.768619 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:53.268497 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:53.768318 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:54.268533 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:54.768555 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:55.267963 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:55.768753 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:56.268259 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:56.769185 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:57.268045 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:57.768862 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:58.268916 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:58.768520 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:59.268987 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:32:59.768451 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:33:00.268339 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:33:00.768065 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:33:01.268676 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:33:01.768439 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:33:02.269185 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:33:02.769506 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:33:03.268773 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:33:03.768361 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:33:04.268042 14042 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0915 06:33:04.769229 14042 kapi.go:107] duration metric: took 2m29.503846329s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0915 06:33:04.771034 14042 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-224201 cluster.
I0915 06:33:04.772373 14042 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0915 06:33:04.773805 14042 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0915 06:33:04.775227 14042 out.go:177] * Enabled addons: ingress-dns, volcano, nvidia-device-plugin, cloud-spanner, storage-provisioner, metrics-server, helm-tiller, inspektor-gadget, yakd, default-storageclass, volumesnapshots, registry, ingress, csi-hostpath-driver, gcp-auth
I0915 06:33:04.776622 14042 addons.go:510] duration metric: took 2m44.890203598s for enable addons: enabled=[ingress-dns volcano nvidia-device-plugin cloud-spanner storage-provisioner metrics-server helm-tiller inspektor-gadget yakd default-storageclass volumesnapshots registry ingress csi-hostpath-driver gcp-auth]
I0915 06:33:04.776669 14042 start.go:246] waiting for cluster config update ...
I0915 06:33:04.776692 14042 start.go:255] writing updated cluster config ...
I0915 06:33:04.777003 14042 ssh_runner.go:195] Run: rm -f paused
I0915 06:33:04.826231 14042 start.go:600] kubectl: 1.31.0, cluster: 1.31.1 (minor skew: 0)
I0915 06:33:04.827899 14042 out.go:177] * Done! kubectl is now configured to use "addons-224201" cluster and "default" namespace by default
==> Docker <==
Sep 15 06:42:24 addons-224201 dockerd[1336]: time="2024-09-15T06:42:24.860536311Z" level=info msg="ignoring event" container=27a22269dc691ecf61992cfe57c82db2844bc8db4c123bdff28317365866d617 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:24 addons-224201 dockerd[1336]: time="2024-09-15T06:42:24.860599645Z" level=info msg="ignoring event" container=667a548984dab7025e9a5c153b196df6dd57e67719569087c55047331035c569 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:24 addons-224201 dockerd[1336]: time="2024-09-15T06:42:24.861127906Z" level=info msg="ignoring event" container=ec61f20e88581a10fa404bd8f1582804f82211653eda3a85b8925316047552af module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:24 addons-224201 dockerd[1336]: time="2024-09-15T06:42:24.864953589Z" level=info msg="ignoring event" container=22f7439f5d7db9094268f8defd8cf70d7c9463dcb76674664953a94f9384f398 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:25 addons-224201 dockerd[1336]: time="2024-09-15T06:42:25.100855875Z" level=info msg="ignoring event" container=8bb525bd710570cdddf8a2422653e83c4f9050068d526d2c41537b90d03e4ce9 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:25 addons-224201 dockerd[1336]: time="2024-09-15T06:42:25.194028276Z" level=info msg="ignoring event" container=2d460f04306ae234e8d5ff3aa4fe3d229132a245133f9907c3b496d5dd8a68e1 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:25 addons-224201 dockerd[1336]: time="2024-09-15T06:42:25.276372147Z" level=info msg="ignoring event" container=c76f287e69a75a8334655a6f423396b24b27f8d03ee36090f683dcd01510ddf0 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:30 addons-224201 dockerd[1336]: time="2024-09-15T06:42:30.811592039Z" level=info msg="ignoring event" container=4d8138e49ea73d6b18c64ad4301ea9ca2e459b704ea6b2440f2f73d55de12557 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:30 addons-224201 dockerd[1336]: time="2024-09-15T06:42:30.814324712Z" level=info msg="ignoring event" container=cad6c2bc1ad6cf04ff2fe33e26be03a884b0f3a5bfe4b9eab6e590c66cb1d9fa module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:30 addons-224201 dockerd[1336]: time="2024-09-15T06:42:30.976094544Z" level=info msg="ignoring event" container=31fe8c9c1e508e45467a7d2f6cc9fb07b9625d5e049cc807af3e574f86887240 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:31 addons-224201 dockerd[1336]: time="2024-09-15T06:42:31.008506910Z" level=info msg="ignoring event" container=a67a569c2ff673d73adcb68709db78a70e765c258e5548ded3ca3e7ed30bfe82 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:32 addons-224201 cri-dockerd[1600]: time="2024-09-15T06:42:32Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/5d7951edc29c2d9d7330fb7abfc74c4d3d72c7e4087d2afe42989b96b4d2e10c/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local us-east4-a.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 15 06:42:32 addons-224201 dockerd[1336]: time="2024-09-15T06:42:32.400564012Z" level=info msg="ignoring event" container=412ba8004bbbf34a590299fc9a1dbc510c28997cb620240676124dd5e85c5535 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:32 addons-224201 dockerd[1336]: time="2024-09-15T06:42:32.459441117Z" level=info msg="ignoring event" container=d7405b5af72bb0c6c630e5896820389136a678e5277a7ebbb835b565f725f623 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:32 addons-224201 cri-dockerd[1600]: time="2024-09-15T06:42:32Z" level=info msg="Stop pulling image docker.io/kicbase/echo-server:1.0: Status: Downloaded newer image for kicbase/echo-server:1.0"
Sep 15 06:42:36 addons-224201 dockerd[1336]: time="2024-09-15T06:42:36.404844864Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=14bb06d3a9bc1246a92853380ffc8a64dba687d3c50afaf1c874cd3e793c8b35
Sep 15 06:42:36 addons-224201 dockerd[1336]: time="2024-09-15T06:42:36.454834708Z" level=info msg="ignoring event" container=14bb06d3a9bc1246a92853380ffc8a64dba687d3c50afaf1c874cd3e793c8b35 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:36 addons-224201 dockerd[1336]: time="2024-09-15T06:42:36.525143995Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 15 06:42:36 addons-224201 dockerd[1336]: time="2024-09-15T06:42:36.526809190Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 15 06:42:36 addons-224201 dockerd[1336]: time="2024-09-15T06:42:36.599388618Z" level=info msg="ignoring event" container=69ff6adeb3ec4bb61553faba5a3d00024bb30259ba1a73e64471117d4ef4f711 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:56 addons-224201 dockerd[1336]: time="2024-09-15T06:42:56.036889867Z" level=info msg="ignoring event" container=c943049539dc6da8b3136156cd09c97de2e73992503bcb6270566d970659e2a6 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:56 addons-224201 dockerd[1336]: time="2024-09-15T06:42:56.501935012Z" level=info msg="ignoring event" container=7dc4f1d47b0bb89644cf3218940b432269642fca88bbfc6988c6aa2d551cd9e5 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:56 addons-224201 dockerd[1336]: time="2024-09-15T06:42:56.575380829Z" level=info msg="ignoring event" container=7ec40d3fcfbf0c288071ea0df72b59c3820e7913be78de24063dd8a482c203a2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:56 addons-224201 dockerd[1336]: time="2024-09-15T06:42:56.665210746Z" level=info msg="ignoring event" container=7f5907c64c99b539f38f9289b22017cd0bc6a6b7fd874b0a78b8adabaccc2f6c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 15 06:42:56 addons-224201 dockerd[1336]: time="2024-09-15T06:42:56.720570453Z" level=info msg="ignoring event" container=9a68d994d60bbbbe69589c42012fbb44d11c7bbd0f101119eb364a4bc65e0950 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
7bda7b72329cd kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 25 seconds ago Running hello-world-app 0 5d7951edc29c2 hello-world-app-55bf9c44b4-hfv5h
d24439de9e42b nginx@sha256:a5127daff3d6f4606be3100a252419bfa84fd6ee5cd74d0feaca1a5068f97dcf 33 seconds ago Running nginx 0 0dde05bf373ee nginx
61c3224c342c0 gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 cfdfd75e1a7eb gcp-auth-89d5ffd79-skd92
c3b3ea8187e39 ce263a8653f9c 11 minutes ago Exited patch 1 79525dd1f424d ingress-nginx-admission-patch-qs5nm
ba10b2491d609 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 c55e38745be57 ingress-nginx-admission-create-vqkr9
d52dfbe09ae26 rancher/local-path-provisioner@sha256:e34c88ae0affb1cdefbb874140d6339d4a27ec4ee420ae8199cd839997b05246 12 minutes ago Running local-path-provisioner 0 01f254fe05107 local-path-provisioner-86d989889c-d9s78
fbf98eb1735ed 6e38f40d628db 12 minutes ago Running storage-provisioner 0 6b78df1f0dc50 storage-provisioner
c34a32ebb6dab c69fa2e9cbf5f 12 minutes ago Running coredns 0 4b6ab499e2cd9 coredns-7c65d6cfc9-q4qbc
9a4313ce9dd66 60c005f310ff3 12 minutes ago Running kube-proxy 0 74652523bdc17 kube-proxy-g8mtm
192b8a4150e3f 175ffd71cce3d 12 minutes ago Running kube-controller-manager 0 96849c8f59029 kube-controller-manager-addons-224201
8354be5e24982 2e96e5913fc06 12 minutes ago Running etcd 0 246fe1b0490ba etcd-addons-224201
d53da43a1283f 6bab7719df100 12 minutes ago Running kube-apiserver 0 ef14a478cff18 kube-apiserver-addons-224201
a297128fb0905 9aa1fad941575 12 minutes ago Running kube-scheduler 0 81d3d5d51094b kube-scheduler-addons-224201
==> coredns [c34a32ebb6da] <==
[INFO] 10.244.0.22:41468 - 35167 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.002383913s
[INFO] 10.244.0.22:42973 - 19331 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.005919882s
[INFO] 10.244.0.22:43055 - 58183 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.067705585s
[INFO] 10.244.0.22:42973 - 25949 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.065524884s
[INFO] 10.244.0.22:41468 - 34531 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.065850504s
[INFO] 10.244.0.22:35536 - 40860 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.065781021s
[INFO] 10.244.0.22:58017 - 29558 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.065974161s
[INFO] 10.244.0.22:46068 - 35488 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.065744229s
[INFO] 10.244.0.22:59784 - 21119 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.065952005s
[INFO] 10.244.0.22:33850 - 16032 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.066103577s
[INFO] 10.244.0.22:43055 - 12473 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000100333s
[INFO] 10.244.0.22:35536 - 22590 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005184429s
[INFO] 10.244.0.22:33850 - 38590 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.006071635s
[INFO] 10.244.0.22:41468 - 44710 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005906025s
[INFO] 10.244.0.22:58017 - 28032 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005649375s
[INFO] 10.244.0.22:33850 - 18397 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000088668s
[INFO] 10.244.0.22:58017 - 43462 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000080216s
[INFO] 10.244.0.22:41468 - 29623 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000068779s
[INFO] 10.244.0.22:35536 - 59829 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000070096s
[INFO] 10.244.0.22:42973 - 23731 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003360063s
[INFO] 10.244.0.22:59784 - 19941 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004755003s
[INFO] 10.244.0.22:46068 - 53126 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005432144s
[INFO] 10.244.0.22:59784 - 48068 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000074335s
[INFO] 10.244.0.22:46068 - 65372 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000043946s
[INFO] 10.244.0.22:42973 - 37626 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000037951s
==> describe nodes <==
Name: addons-224201
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-224201
kubernetes.io/os=linux
minikube.k8s.io/commit=7a3ca67a20528f5dabbb456e8e4ce542b58ef23a
minikube.k8s.io/name=addons-224201
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_15T06_30_15_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-224201
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Sun, 15 Sep 2024 06:30:12 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-224201
AcquireTime: <unset>
RenewTime: Sun, 15 Sep 2024 06:42:49 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Sun, 15 Sep 2024 06:42:49 +0000 Sun, 15 Sep 2024 06:30:10 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Sun, 15 Sep 2024 06:42:49 +0000 Sun, 15 Sep 2024 06:30:10 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Sun, 15 Sep 2024 06:42:49 +0000 Sun, 15 Sep 2024 06:30:10 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Sun, 15 Sep 2024 06:42:49 +0000 Sun, 15 Sep 2024 06:30:12 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-224201
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
System Info:
Machine ID: ab4f1039ebe14feea123508f518077b7
System UUID: 94c9c2ca-65ef-4362-903f-e485902d9a84
Boot ID: 3a2c343e-04b3-4276-bce8-4c4fba0d38ee
Kernel Version: 5.15.0-1068-gcp
OS Image: Ubuntu 22.04.4 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.2.1
Kubelet Version: v1.31.1
Kube-Proxy Version: v1.31.1
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (12 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m14s
default hello-world-app-55bf9c44b4-hfv5h 0 (0%) 0 (0%) 0 (0%) 0 (0%) 26s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 35s
gcp-auth gcp-auth-89d5ffd79-skd92 0 (0%) 0 (0%) 0 (0%) 0 (0%) 11m
kube-system coredns-7c65d6cfc9-q4qbc 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-224201 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-224201 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-224201 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-g8mtm 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-224201 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
local-path-storage local-path-provisioner-86d989889c-d9s78 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-224201 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-224201 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-224201 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-224201 event: Registered Node addons-224201 in Controller
Normal CIDRAssignmentFailed 12m cidrAllocator Node addons-224201 status is now: CIDRAssignmentFailed
==> dmesg <==
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 7e 0b 08 ef 16 19 08 06
[ +1.216227] IPv4: martian source 10.244.0.1 from 10.244.0.18, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 86 19 e7 0e 08 f2 08 06
[ +7.309979] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 5a a0 73 4a 6c 47 08 06
[ +2.868715] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 36 5e 36 81 cd 04 08 06
[ +0.270770] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 0e b4 b8 26 7f 6e 08 06
[ +0.432526] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff b2 30 c3 1b 0f c3 08 06
[Sep15 06:32] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 3e 9e 87 a7 41 a4 08 06
[ +0.057455] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 7e 38 88 07 d8 bc 08 06
[Sep15 06:33] IPv4: martian source 10.244.0.1 from 10.244.0.26, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff c6 ad 81 16 ec 97 08 06
[ +0.000435] IPv4: martian source 10.244.0.26 from 10.244.0.2, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff 66 02 6a 75 a0 7d 08 06
[Sep15 06:42] IPv4: martian source 10.244.0.1 from 10.244.0.35, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff b6 1a 5c 15 95 e5 08 06
[ +21.166418] IPv4: martian source 10.244.0.37 from 10.244.0.22, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 5a a0 73 4a 6c 47 08 06
[ +1.454901] IPv4: martian source 10.244.0.22 from 10.244.0.2, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 66 02 6a 75 a0 7d 08 06
==> etcd [8354be5e2498] <==
{"level":"info","ts":"2024-09-15T06:30:10.570932Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-15T06:30:10.570949Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-09-15T06:30:10.570962Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-15T06:30:10.572565Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-15T06:30:10.573145Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-15T06:30:10.573139Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-224201 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-09-15T06:30:10.573175Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-15T06:30:10.573333Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-15T06:30:10.573407Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-15T06:30:10.573519Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-15T06:30:10.573606Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-15T06:30:10.573625Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-15T06:30:10.574313Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-15T06:30:10.574633Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-15T06:30:10.575390Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-15T06:30:10.575433Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"warn","ts":"2024-09-15T06:30:30.777407Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"105.076325ms","expected-duration":"100ms","prefix":"","request":"header:<ID:8128031906263597089 username:\"kube-apiserver-etcd-client\" auth_revision:1 > txn:<compare:<target:MOD key:\"/registry/apiregistration.k8s.io/apiservices/v1alpha1.batch.volcano.sh\" mod_revision:0 > success:<request_put:<key:\"/registry/apiregistration.k8s.io/apiservices/v1alpha1.batch.volcano.sh\" value_size:893 >> failure:<>>","response":"size:16"}
{"level":"info","ts":"2024-09-15T06:30:30.777615Z","caller":"traceutil/trace.go:171","msg":"trace[1179557176] transaction","detail":"{read_only:false; response_revision:763; number_of_response:1; }","duration":"107.7824ms","start":"2024-09-15T06:30:30.669818Z","end":"2024-09-15T06:30:30.777601Z","steps":["trace[1179557176] 'compare' (duration: 104.967116ms)"],"step_count":1}
{"level":"info","ts":"2024-09-15T06:30:38.386110Z","caller":"traceutil/trace.go:171","msg":"trace[4864068] linearizableReadLoop","detail":"{readStateIndex:989; appliedIndex:988; }","duration":"118.744756ms","start":"2024-09-15T06:30:38.267351Z","end":"2024-09-15T06:30:38.386095Z","steps":["trace[4864068] 'read index received' (duration: 56.5526ms)","trace[4864068] 'applied index is now lower than readState.Index' (duration: 62.191452ms)"],"step_count":2}
{"level":"info","ts":"2024-09-15T06:30:38.386158Z","caller":"traceutil/trace.go:171","msg":"trace[88716142] transaction","detail":"{read_only:false; response_revision:971; number_of_response:1; }","duration":"122.321859ms","start":"2024-09-15T06:30:38.263819Z","end":"2024-09-15T06:30:38.386140Z","steps":["trace[88716142] 'process raft request' (duration: 60.115194ms)","trace[88716142] 'compare' (duration: 62.099191ms)"],"step_count":2}
{"level":"warn","ts":"2024-09-15T06:30:38.386273Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"118.879859ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods\" limit:1 ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-09-15T06:30:38.386313Z","caller":"traceutil/trace.go:171","msg":"trace[406536097] range","detail":"{range_begin:/registry/pods; range_end:; response_count:0; response_revision:971; }","duration":"118.957356ms","start":"2024-09-15T06:30:38.267347Z","end":"2024-09-15T06:30:38.386305Z","steps":["trace[406536097] 'agreement among raft nodes before linearized reading' (duration: 118.802358ms)"],"step_count":1}
{"level":"info","ts":"2024-09-15T06:40:11.175097Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1891}
{"level":"info","ts":"2024-09-15T06:40:11.198512Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1891,"took":"22.892942ms","hash":2140157512,"current-db-size-bytes":9232384,"current-db-size":"9.2 MB","current-db-size-in-use-bytes":5017600,"current-db-size-in-use":"5.0 MB"}
{"level":"info","ts":"2024-09-15T06:40:11.198548Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":2140157512,"revision":1891,"compact-revision":-1}
==> gcp-auth [61c3224c342c] <==
2024/09/15 06:33:43 Ready to write response ...
2024/09/15 06:41:46 Ready to marshal response ...
2024/09/15 06:41:46 Ready to write response ...
2024/09/15 06:41:46 Ready to marshal response ...
2024/09/15 06:41:46 Ready to write response ...
2024/09/15 06:41:48 Ready to marshal response ...
2024/09/15 06:41:48 Ready to write response ...
2024/09/15 06:41:53 Ready to marshal response ...
2024/09/15 06:41:53 Ready to write response ...
2024/09/15 06:41:55 Ready to marshal response ...
2024/09/15 06:41:55 Ready to write response ...
2024/09/15 06:42:06 Ready to marshal response ...
2024/09/15 06:42:06 Ready to write response ...
2024/09/15 06:42:06 Ready to marshal response ...
2024/09/15 06:42:06 Ready to write response ...
2024/09/15 06:42:06 Ready to marshal response ...
2024/09/15 06:42:06 Ready to write response ...
2024/09/15 06:42:08 Ready to marshal response ...
2024/09/15 06:42:08 Ready to write response ...
2024/09/15 06:42:16 Ready to marshal response ...
2024/09/15 06:42:16 Ready to write response ...
2024/09/15 06:42:22 Ready to marshal response ...
2024/09/15 06:42:22 Ready to write response ...
2024/09/15 06:42:31 Ready to marshal response ...
2024/09/15 06:42:31 Ready to write response ...
==> kernel <==
06:42:57 up 25 min, 0 users, load average: 0.31, 0.42, 0.44
Linux addons-224201 5.15.0-1068-gcp #76~20.04.1-Ubuntu SMP Tue Aug 20 15:52:45 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.4 LTS"
==> kube-apiserver [d53da43a1283] <==
W0915 06:33:35.370027 1 cacher.go:171] Terminating all watchers from cacher numatopologies.nodeinfo.volcano.sh
W0915 06:33:35.492074 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0915 06:33:35.770072 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0915 06:33:36.168102 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0915 06:41:55.135347 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0915 06:42:06.330487 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.103.157.79"}
I0915 06:42:19.201649 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0915 06:42:20.218209 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0915 06:42:22.154001 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
I0915 06:42:22.307205 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.103.135.177"}
I0915 06:42:30.733783 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0915 06:42:30.733833 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0915 06:42:30.745098 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0915 06:42:30.745145 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0915 06:42:30.745857 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0915 06:42:30.745895 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0915 06:42:30.756442 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0915 06:42:30.756477 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0915 06:42:30.767852 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0915 06:42:30.767889 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
W0915 06:42:31.746352 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
W0915 06:42:31.768490 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
I0915 06:42:31.782154 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.108.32.40"}
W0915 06:42:31.876961 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
I0915 06:42:51.233212 1 controller.go:129] OpenAPI AggregationController: action for item v1beta1.metrics.k8s.io: Nothing (removed from the queue).
==> kube-controller-manager [192b8a4150e3] <==
W0915 06:42:39.276335 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 06:42:39.276368 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0915 06:42:41.049492 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 06:42:41.049527 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0915 06:42:41.082148 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 06:42:41.082181 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0915 06:42:43.350890 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="ingress-nginx"
W0915 06:42:45.231351 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 06:42:45.231386 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0915 06:42:45.737542 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 06:42:45.737577 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0915 06:42:47.930960 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 06:42:47.931004 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0915 06:42:49.110410 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-224201"
W0915 06:42:49.480157 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 06:42:49.480199 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0915 06:42:49.526408 1 shared_informer.go:313] Waiting for caches to sync for resource quota
I0915 06:42:49.526442 1 shared_informer.go:320] Caches are synced for resource quota
I0915 06:42:49.846129 1 shared_informer.go:313] Waiting for caches to sync for garbage collector
I0915 06:42:49.846166 1 shared_informer.go:320] Caches are synced for garbage collector
W0915 06:42:50.902238 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 06:42:50.902280 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0915 06:42:56.462421 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-66c9cd494c" duration="6.702µs"
W0915 06:42:57.413408 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0915 06:42:57.413455 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
==> kube-proxy [9a4313ce9dd6] <==
I0915 06:30:23.586185 1 server_linux.go:66] "Using iptables proxy"
I0915 06:30:24.177645 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0915 06:30:24.177719 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0915 06:30:24.468723 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0915 06:30:24.468804 1 server_linux.go:169] "Using iptables Proxier"
I0915 06:30:24.480776 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0915 06:30:24.481064 1 server.go:483] "Version info" version="v1.31.1"
I0915 06:30:24.481083 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0915 06:30:24.482677 1 config.go:199] "Starting service config controller"
I0915 06:30:24.482690 1 shared_informer.go:313] Waiting for caches to sync for service config
I0915 06:30:24.482708 1 config.go:105] "Starting endpoint slice config controller"
I0915 06:30:24.482712 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0915 06:30:24.482993 1 config.go:328] "Starting node config controller"
I0915 06:30:24.482998 1 shared_informer.go:313] Waiting for caches to sync for node config
I0915 06:30:24.658157 1 shared_informer.go:320] Caches are synced for node config
I0915 06:30:24.658171 1 shared_informer.go:320] Caches are synced for service config
I0915 06:30:24.658187 1 shared_informer.go:320] Caches are synced for endpoint slice config
==> kube-scheduler [a297128fb090] <==
W0915 06:30:12.282589 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0915 06:30:12.282626 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0915 06:30:12.282321 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0915 06:30:12.282673 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0915 06:30:12.282400 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0915 06:30:12.282712 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0915 06:30:12.282404 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0915 06:30:12.282745 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0915 06:30:12.282404 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope
E0915 06:30:12.282774 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0915 06:30:12.282488 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0915 06:30:12.282798 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0915 06:30:12.282507 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0915 06:30:12.282819 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0915 06:30:13.094718 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope
E0915 06:30:13.094768 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes is forbidden: User \"system:kube-scheduler\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0915 06:30:13.125227 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0915 06:30:13.125263 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0915 06:30:13.143750 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0915 06:30:13.143790 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0915 06:30:13.326910 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0915 06:30:13.326951 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"statefulsets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0915 06:30:13.458554 1 reflector.go:561] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0915 06:30:13.458601 1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
I0915 06:30:15.279183 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 15 06:42:36 addons-224201 kubelet[2443]: I0915 06:42:36.881936 2443 reconciler_common.go:288] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1dadbfde-cd51-46ad-92ba-ca815c6f01b0-webhook-cert\") on node \"addons-224201\" DevicePath \"\""
Sep 15 06:42:38 addons-224201 kubelet[2443]: I0915 06:42:38.478662 2443 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dadbfde-cd51-46ad-92ba-ca815c6f01b0" path="/var/lib/kubelet/pods/1dadbfde-cd51-46ad-92ba-ca815c6f01b0/volumes"
Sep 15 06:42:46 addons-224201 kubelet[2443]: E0915 06:42:46.471474 2443 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="9a8dcbeb-7d45-426d-b462-9bc8f185c665"
Sep 15 06:42:48 addons-224201 kubelet[2443]: E0915 06:42:48.471528 2443 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-test\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox\\\"\"" pod="default/registry-test" podUID="e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd"
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.189122 2443 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd-gcp-creds\") pod \"e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd\" (UID: \"e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd\") "
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.189178 2443 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfqg9\" (UniqueName: \"kubernetes.io/projected/e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd-kube-api-access-hfqg9\") pod \"e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd\" (UID: \"e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd\") "
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.189221 2443 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd" (UID: "e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.190891 2443 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd-kube-api-access-hfqg9" (OuterVolumeSpecName: "kube-api-access-hfqg9") pod "e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd" (UID: "e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd"). InnerVolumeSpecName "kube-api-access-hfqg9". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.290202 2443 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-hfqg9\" (UniqueName: \"kubernetes.io/projected/e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd-kube-api-access-hfqg9\") on node \"addons-224201\" DevicePath \"\""
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.290237 2443 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/e7120fcf-4bc1-4cba-9dff-d6c63f0f7abd-gcp-creds\") on node \"addons-224201\" DevicePath \"\""
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.894391 2443 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sncm2\" (UniqueName: \"kubernetes.io/projected/48d3ee3d-2a77-4bbb-9492-d6cb8accc5e0-kube-api-access-sncm2\") pod \"48d3ee3d-2a77-4bbb-9492-d6cb8accc5e0\" (UID: \"48d3ee3d-2a77-4bbb-9492-d6cb8accc5e0\") "
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.894453 2443 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jznx\" (UniqueName: \"kubernetes.io/projected/8343911e-7ef0-4555-a6d5-acefa24b2ae3-kube-api-access-9jznx\") pod \"8343911e-7ef0-4555-a6d5-acefa24b2ae3\" (UID: \"8343911e-7ef0-4555-a6d5-acefa24b2ae3\") "
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.896194 2443 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48d3ee3d-2a77-4bbb-9492-d6cb8accc5e0-kube-api-access-sncm2" (OuterVolumeSpecName: "kube-api-access-sncm2") pod "48d3ee3d-2a77-4bbb-9492-d6cb8accc5e0" (UID: "48d3ee3d-2a77-4bbb-9492-d6cb8accc5e0"). InnerVolumeSpecName "kube-api-access-sncm2". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.896401 2443 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8343911e-7ef0-4555-a6d5-acefa24b2ae3-kube-api-access-9jznx" (OuterVolumeSpecName: "kube-api-access-9jznx") pod "8343911e-7ef0-4555-a6d5-acefa24b2ae3" (UID: "8343911e-7ef0-4555-a6d5-acefa24b2ae3"). InnerVolumeSpecName "kube-api-access-9jznx". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.951022 2443 scope.go:117] "RemoveContainer" containerID="7ec40d3fcfbf0c288071ea0df72b59c3820e7913be78de24063dd8a482c203a2"
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.966195 2443 scope.go:117] "RemoveContainer" containerID="7ec40d3fcfbf0c288071ea0df72b59c3820e7913be78de24063dd8a482c203a2"
Sep 15 06:42:56 addons-224201 kubelet[2443]: E0915 06:42:56.966919 2443 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 7ec40d3fcfbf0c288071ea0df72b59c3820e7913be78de24063dd8a482c203a2" containerID="7ec40d3fcfbf0c288071ea0df72b59c3820e7913be78de24063dd8a482c203a2"
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.966964 2443 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"7ec40d3fcfbf0c288071ea0df72b59c3820e7913be78de24063dd8a482c203a2"} err="failed to get container status \"7ec40d3fcfbf0c288071ea0df72b59c3820e7913be78de24063dd8a482c203a2\": rpc error: code = Unknown desc = Error response from daemon: No such container: 7ec40d3fcfbf0c288071ea0df72b59c3820e7913be78de24063dd8a482c203a2"
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.966991 2443 scope.go:117] "RemoveContainer" containerID="7dc4f1d47b0bb89644cf3218940b432269642fca88bbfc6988c6aa2d551cd9e5"
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.984341 2443 scope.go:117] "RemoveContainer" containerID="7dc4f1d47b0bb89644cf3218940b432269642fca88bbfc6988c6aa2d551cd9e5"
Sep 15 06:42:56 addons-224201 kubelet[2443]: E0915 06:42:56.987347 2443 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 7dc4f1d47b0bb89644cf3218940b432269642fca88bbfc6988c6aa2d551cd9e5" containerID="7dc4f1d47b0bb89644cf3218940b432269642fca88bbfc6988c6aa2d551cd9e5"
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.987388 2443 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"7dc4f1d47b0bb89644cf3218940b432269642fca88bbfc6988c6aa2d551cd9e5"} err="failed to get container status \"7dc4f1d47b0bb89644cf3218940b432269642fca88bbfc6988c6aa2d551cd9e5\": rpc error: code = Unknown desc = Error response from daemon: No such container: 7dc4f1d47b0bb89644cf3218940b432269642fca88bbfc6988c6aa2d551cd9e5"
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.995608 2443 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-sncm2\" (UniqueName: \"kubernetes.io/projected/48d3ee3d-2a77-4bbb-9492-d6cb8accc5e0-kube-api-access-sncm2\") on node \"addons-224201\" DevicePath \"\""
Sep 15 06:42:56 addons-224201 kubelet[2443]: I0915 06:42:56.995639 2443 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-9jznx\" (UniqueName: \"kubernetes.io/projected/8343911e-7ef0-4555-a6d5-acefa24b2ae3-kube-api-access-9jznx\") on node \"addons-224201\" DevicePath \"\""
Sep 15 06:42:57 addons-224201 kubelet[2443]: E0915 06:42:57.472118 2443 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"busybox\" with ImagePullBackOff: \"Back-off pulling image \\\"gcr.io/k8s-minikube/busybox:1.28.4-glibc\\\"\"" pod="default/busybox" podUID="9a8dcbeb-7d45-426d-b462-9bc8f185c665"
==> storage-provisioner [fbf98eb1735e] <==
I0915 06:30:27.271580 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0915 06:30:27.283057 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0915 06:30:27.283113 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0915 06:30:27.367533 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0915 06:30:27.367725 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-224201_25bfd251-0330-41e1-b038-f9b05a381621!
I0915 06:30:27.369358 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"c25c5532-ac94-44d0-8aa4-1200dfa9de8e", APIVersion:"v1", ResourceVersion:"599", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-224201_25bfd251-0330-41e1-b038-f9b05a381621 became leader
I0915 06:30:27.468340 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-224201_25bfd251-0330-41e1-b038-f9b05a381621!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-224201 -n addons-224201
helpers_test.go:261: (dbg) Run: kubectl --context addons-224201 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-224201 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-224201 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-224201/192.168.49.2
Start Time: Sun, 15 Sep 2024 06:33:43 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.28
IPs:
IP: 10.244.0.28
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-l4qtv (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-l4qtv:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m15s default-scheduler Successfully assigned default/busybox to addons-224201
Normal Pulling 7m41s (x4 over 9m14s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m41s (x4 over 9m14s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m41s (x4 over 9m14s) kubelet Error: ErrImagePull
Warning Failed 7m30s (x6 over 9m14s) kubelet Error: ImagePullBackOff
Normal BackOff 4m13s (x20 over 9m14s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (72.31s)