=== RUN TestAddons/parallel/CSI
=== PAUSE TestAddons/parallel/CSI
=== CONT TestAddons/parallel/CSI
addons_test.go:563: csi-hostpath-driver pods stabilized in 4.692871ms
addons_test.go:566: (dbg) Run: kubectl --context addons-246651 create -f testdata/csi-hostpath-driver/pvc.yaml
addons_test.go:571: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pvc "hpvc" in namespace "default" ...
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc -o jsonpath={.status.phase} -n default
addons_test.go:576: (dbg) Run: kubectl --context addons-246651 create -f testdata/csi-hostpath-driver/pv-pod.yaml
addons_test.go:581: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pods matching "app=task-pv-pod" in namespace "default" ...
helpers_test.go:344: "task-pv-pod" [4c10037d-7648-4a2f-b362-c3ccebe84c6f] Pending
helpers_test.go:344: "task-pv-pod" [4c10037d-7648-4a2f-b362-c3ccebe84c6f] Pending / Ready:ContainersNotReady (containers with unready status: [task-pv-container]) / ContainersReady:ContainersNotReady (containers with unready status: [task-pv-container])
helpers_test.go:344: "task-pv-pod" [4c10037d-7648-4a2f-b362-c3ccebe84c6f] Running
addons_test.go:581: (dbg) TestAddons/parallel/CSI: app=task-pv-pod healthy within 12.004253522s
addons_test.go:586: (dbg) Run: kubectl --context addons-246651 create -f testdata/csi-hostpath-driver/snapshot.yaml
addons_test.go:591: (dbg) TestAddons/parallel/CSI: waiting 6m0s for volume snapshot "new-snapshot-demo" in namespace "default" ...
helpers_test.go:419: (dbg) Run: kubectl --context addons-246651 get volumesnapshot new-snapshot-demo -o jsonpath={.status.readyToUse} -n default
helpers_test.go:419: (dbg) Run: kubectl --context addons-246651 get volumesnapshot new-snapshot-demo -o jsonpath={.status.readyToUse} -n default
addons_test.go:596: (dbg) Run: kubectl --context addons-246651 delete pod task-pv-pod
addons_test.go:602: (dbg) Run: kubectl --context addons-246651 delete pvc hpvc
addons_test.go:608: (dbg) Run: kubectl --context addons-246651 create -f testdata/csi-hostpath-driver/pvc-restore.yaml
addons_test.go:613: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pvc "hpvc-restore" in namespace "default" ...
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
helpers_test.go:394: (dbg) Run: kubectl --context addons-246651 get pvc hpvc-restore -o jsonpath={.status.phase} -n default
addons_test.go:618: (dbg) Run: kubectl --context addons-246651 create -f testdata/csi-hostpath-driver/pv-pod-restore.yaml
addons_test.go:618: (dbg) Non-zero exit: kubectl --context addons-246651 create -f testdata/csi-hostpath-driver/pv-pod-restore.yaml: exit status 1 (189.906026ms)
** stderr **
Error from server (InternalError): error when creating "testdata/csi-hostpath-driver/pv-pod-restore.yaml": Internal error occurred: failed calling webhook "validatepod.volcano.sh": failed to call webhook: Post "https://volcano-admission-service.volcano-system.svc:443/pods/validate?timeout=10s": service "volcano-admission-service" not found
** /stderr **
addons_test.go:620: creating pod with kubectl --context addons-246651 create -f testdata/csi-hostpath-driver/pv-pod-restore.yaml failed: exit status 1
addons_test.go:623: (dbg) TestAddons/parallel/CSI: waiting 6m0s for pods matching "app=task-pv-pod-restore" in namespace "default" ...
helpers_test.go:329: TestAddons/parallel/CSI: WARNING: pod list for "default" "app=task-pv-pod-restore" returned: client rate limiter Wait returned an error: context deadline exceeded
addons_test.go:623: ***** TestAddons/parallel/CSI: pod "app=task-pv-pod-restore" failed to start within 6m0s: context deadline exceeded ****
addons_test.go:623: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-246651 -n addons-246651
addons_test.go:623: TestAddons/parallel/CSI: showing logs for failed pods as of 2024-07-17 00:31:27.182814248 +0000 UTC m=+656.729973354
addons_test.go:624: failed waiting for pod task-pv-pod-restore: app=task-pv-pod-restore within 6m0s: context deadline exceeded
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/CSI]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-246651
helpers_test.go:235: (dbg) docker inspect addons-246651:
-- stdout --
[
{
"Id": "6cb175652c6790236be8e04e03a61f524a7ba87389827a0b05962825cf1fd469",
"Created": "2024-07-17T00:21:32.134891001Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 12862,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-07-17T00:21:32.257570836Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:b90fcd82d9a0f97666ccbedd0bec36ffa6ae451ed5f5fff480c00361af0818c6",
"ResolvConfPath": "/var/lib/docker/containers/6cb175652c6790236be8e04e03a61f524a7ba87389827a0b05962825cf1fd469/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/6cb175652c6790236be8e04e03a61f524a7ba87389827a0b05962825cf1fd469/hostname",
"HostsPath": "/var/lib/docker/containers/6cb175652c6790236be8e04e03a61f524a7ba87389827a0b05962825cf1fd469/hosts",
"LogPath": "/var/lib/docker/containers/6cb175652c6790236be8e04e03a61f524a7ba87389827a0b05962825cf1fd469/6cb175652c6790236be8e04e03a61f524a7ba87389827a0b05962825cf1fd469-json.log",
"Name": "/addons-246651",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"addons-246651:/var",
"/lib/modules:/lib/modules:ro"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-246651",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/f92667ac81d592ad0d6416be061da6112107d327467039d131aa2173576502bf-init/diff:/var/lib/docker/overlay2/7d8df8d1f86f633603e9fb45a180f0e54bf8e3b246527d0b435ec84b095e56c3/diff",
"MergedDir": "/var/lib/docker/overlay2/f92667ac81d592ad0d6416be061da6112107d327467039d131aa2173576502bf/merged",
"UpperDir": "/var/lib/docker/overlay2/f92667ac81d592ad0d6416be061da6112107d327467039d131aa2173576502bf/diff",
"WorkDir": "/var/lib/docker/overlay2/f92667ac81d592ad0d6416be061da6112107d327467039d131aa2173576502bf/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "volume",
"Name": "addons-246651",
"Source": "/var/lib/docker/volumes/addons-246651/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
},
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
}
],
"Config": {
"Hostname": "addons-246651",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-246651",
"name.minikube.sigs.k8s.io": "addons-246651",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "17dc5c97828119675cc1597dafc1c62ddd0c39a1ecb28b3521f9d86ed3f34ff5",
"SandboxKey": "/var/run/docker/netns/17dc5c978281",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-246651": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "3defc5bdc56ba4905b96b250e0ea6a959d80f3f0493891074bbcc5e38b50af7e",
"EndpointID": "6e72f5292fb8ab3d2a4cb36321e594bacab04dfa81dbfa5bbddf8f5b0ccd7669",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-246651",
"6cb175652c67"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-246651 -n addons-246651
helpers_test.go:244: <<< TestAddons/parallel/CSI FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/CSI]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-246651 logs -n 25
helpers_test.go:252: TestAddons/parallel/CSI logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-only-671576 | download-only-671576 | jenkins | v1.33.1 | 17 Jul 24 00:21 UTC | 17 Jul 24 00:21 UTC |
| delete | -p download-only-673382 | download-only-673382 | jenkins | v1.33.1 | 17 Jul 24 00:21 UTC | 17 Jul 24 00:21 UTC |
| delete | -p download-only-966512 | download-only-966512 | jenkins | v1.33.1 | 17 Jul 24 00:21 UTC | 17 Jul 24 00:21 UTC |
| start | --download-only -p | download-docker-523526 | jenkins | v1.33.1 | 17 Jul 24 00:21 UTC | |
| | download-docker-523526 | | | | | |
| | --alsologtostderr | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p download-docker-523526 | download-docker-523526 | jenkins | v1.33.1 | 17 Jul 24 00:21 UTC | 17 Jul 24 00:21 UTC |
| start | --download-only -p | binary-mirror-452108 | jenkins | v1.33.1 | 17 Jul 24 00:21 UTC | |
| | binary-mirror-452108 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:43789 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-452108 | binary-mirror-452108 | jenkins | v1.33.1 | 17 Jul 24 00:21 UTC | 17 Jul 24 00:21 UTC |
| addons | disable dashboard -p | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:21 UTC | |
| | addons-246651 | | | | | |
| addons | enable dashboard -p | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:21 UTC | |
| | addons-246651 | | | | | |
| start | -p addons-246651 --wait=true | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:21 UTC | 17 Jul 24 00:24 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | enable headlamp | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:24 UTC | 17 Jul 24 00:24 UTC |
| | -p addons-246651 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-246651 addons | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:24 UTC | 17 Jul 24 00:24 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-246651 addons disable | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:24 UTC | 17 Jul 24 00:24 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| ip | addons-246651 ip | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| addons | addons-246651 addons disable | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| | -p addons-246651 | | | | | |
| addons | disable cloud-spanner -p | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| | addons-246651 | | | | | |
| ssh | addons-246651 ssh cat | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| | /opt/local-path-provisioner/pvc-ebf19b2d-b2d8-4fd9-99a3-42f79b8e2ec6_default_test-pvc/file1 | | | | | |
| addons | addons-246651 addons disable | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| ssh | addons-246651 ssh curl -s | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-246651 ip | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| addons | addons-246651 addons disable | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | addons-246651 addons disable | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-246651 addons disable | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-246651 | jenkins | v1.33.1 | 17 Jul 24 00:25 UTC | 17 Jul 24 00:25 UTC |
| | addons-246651 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/07/17 00:21:10
Running on machine: ubuntu-20-agent-8
Binary: Built with gc go1.22.5 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0717 00:21:10.078763 12110 out.go:291] Setting OutFile to fd 1 ...
I0717 00:21:10.078855 12110 out.go:338] TERM=,COLORTERM=, which probably does not support color
I0717 00:21:10.078861 12110 out.go:304] Setting ErrFile to fd 2...
I0717 00:21:10.078865 12110 out.go:338] TERM=,COLORTERM=, which probably does not support color
I0717 00:21:10.079059 12110 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19264-3553/.minikube/bin
I0717 00:21:10.079620 12110 out.go:298] Setting JSON to false
I0717 00:21:10.080385 12110 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-8","uptime":168,"bootTime":1721175502,"procs":180,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1062-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0717 00:21:10.080438 12110 start.go:139] virtualization: kvm guest
I0717 00:21:10.082843 12110 out.go:177] * [addons-246651] minikube v1.33.1 on Ubuntu 20.04 (kvm/amd64)
I0717 00:21:10.084325 12110 notify.go:220] Checking for updates...
I0717 00:21:10.084331 12110 out.go:177] - MINIKUBE_LOCATION=19264
I0717 00:21:10.085454 12110 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0717 00:21:10.086846 12110 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19264-3553/kubeconfig
I0717 00:21:10.087920 12110 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19264-3553/.minikube
I0717 00:21:10.089220 12110 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0717 00:21:10.090324 12110 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0717 00:21:10.091614 12110 driver.go:392] Setting default libvirt URI to qemu:///system
I0717 00:21:10.112376 12110 docker.go:123] docker version: linux-27.0.3:Docker Engine - Community
I0717 00:21:10.112459 12110 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0717 00:21:10.160233 12110 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-07-17 00:21:10.151071396 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1062-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647951872 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-8 Labels:[] ExperimentalBuild:false ServerVersion:27.0.3 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:ae71819c4f5e67bb4d5ae76a6b735f29cc25774e Expected:ae71819c4f5e67bb4d5ae76a6b735f29cc25774e} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErr
ors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.15.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.28.1] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0717 00:21:10.160377 12110 docker.go:307] overlay module found
I0717 00:21:10.162114 12110 out.go:177] * Using the docker driver based on user configuration
I0717 00:21:10.163420 12110 start.go:297] selected driver: docker
I0717 00:21:10.163445 12110 start.go:901] validating driver "docker" against <nil>
I0717 00:21:10.163461 12110 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0717 00:21:10.164472 12110 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0717 00:21:10.212439 12110 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-07-17 00:21:10.204449562 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1062-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647951872 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-8 Labels:[] ExperimentalBuild:false ServerVersion:27.0.3 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:ae71819c4f5e67bb4d5ae76a6b735f29cc25774e Expected:ae71819c4f5e67bb4d5ae76a6b735f29cc25774e} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErr
ors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.15.1] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.28.1] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0717 00:21:10.212650 12110 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0717 00:21:10.212928 12110 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0717 00:21:10.214553 12110 out.go:177] * Using Docker driver with root privileges
I0717 00:21:10.216223 12110 cni.go:84] Creating CNI manager for ""
I0717 00:21:10.216255 12110 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0717 00:21:10.216269 12110 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0717 00:21:10.216359 12110 start.go:340] cluster config:
{Name:addons-246651 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.2 ClusterName:addons-246651 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime
:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.30.2 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock:
SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0717 00:21:10.217824 12110 out.go:177] * Starting "addons-246651" primary control-plane node in "addons-246651" cluster
I0717 00:21:10.218878 12110 cache.go:121] Beginning downloading kic base image for docker with docker
I0717 00:21:10.220044 12110 out.go:177] * Pulling base image v0.0.44-1721146479-19264 ...
I0717 00:21:10.221138 12110 preload.go:131] Checking if preload exists for k8s version v1.30.2 and runtime docker
I0717 00:21:10.221177 12110 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19264-3553/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.30.2-docker-overlay2-amd64.tar.lz4
I0717 00:21:10.221188 12110 cache.go:56] Caching tarball of preloaded images
I0717 00:21:10.221252 12110 image.go:79] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e in local docker daemon
I0717 00:21:10.221272 12110 preload.go:172] Found /home/jenkins/minikube-integration/19264-3553/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.30.2-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0717 00:21:10.221397 12110 cache.go:59] Finished verifying existence of preloaded tar for v1.30.2 on docker
I0717 00:21:10.221791 12110 profile.go:143] Saving config to /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/config.json ...
I0717 00:21:10.221817 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/config.json: {Name:mk335f18c5abf31e994eb4919fb1b9c4e14aa1cc Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:10.236694 12110 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e to local cache
I0717 00:21:10.236813 12110 image.go:63] Checking for gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e in local cache directory
I0717 00:21:10.236831 12110 image.go:66] Found gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e in local cache directory, skipping pull
I0717 00:21:10.236837 12110 image.go:135] gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e exists in cache, skipping pull
I0717 00:21:10.236848 12110 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e as a tarball
I0717 00:21:10.236858 12110 cache.go:162] Loading gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e from local cache
I0717 00:21:22.253414 12110 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e from cached tarball
I0717 00:21:22.253448 12110 cache.go:194] Successfully downloaded all kic artifacts
I0717 00:21:22.253502 12110 start.go:360] acquireMachinesLock for addons-246651: {Name:mk20c165273aeb30b560598bba6d1d9c0260bacb Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0717 00:21:22.253608 12110 start.go:364] duration metric: took 85.272µs to acquireMachinesLock for "addons-246651"
I0717 00:21:22.253634 12110 start.go:93] Provisioning new machine with config: &{Name:addons-246651 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.2 ClusterName:addons-246651 Namespace:default APIServerHAVIP: APIServerName:min
ikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.30.2 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false Cust
omQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.30.2 ContainerRuntime:docker ControlPlane:true Worker:true}
I0717 00:21:22.253728 12110 start.go:125] createHost starting for "" (driver="docker")
I0717 00:21:22.255509 12110 out.go:204] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0717 00:21:22.255727 12110 start.go:159] libmachine.API.Create for "addons-246651" (driver="docker")
I0717 00:21:22.255756 12110 client.go:168] LocalClient.Create starting
I0717 00:21:22.255868 12110 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19264-3553/.minikube/certs/ca.pem
I0717 00:21:22.482632 12110 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19264-3553/.minikube/certs/cert.pem
I0717 00:21:22.970668 12110 cli_runner.go:164] Run: docker network inspect addons-246651 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0717 00:21:22.985818 12110 cli_runner.go:211] docker network inspect addons-246651 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0717 00:21:22.985890 12110 network_create.go:284] running [docker network inspect addons-246651] to gather additional debugging logs...
I0717 00:21:22.985910 12110 cli_runner.go:164] Run: docker network inspect addons-246651
W0717 00:21:23.002273 12110 cli_runner.go:211] docker network inspect addons-246651 returned with exit code 1
I0717 00:21:23.002300 12110 network_create.go:287] error running [docker network inspect addons-246651]: docker network inspect addons-246651: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-246651 not found
I0717 00:21:23.002315 12110 network_create.go:289] output of [docker network inspect addons-246651]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-246651 not found
** /stderr **
I0717 00:21:23.002401 12110 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0717 00:21:23.018675 12110 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001abc490}
I0717 00:21:23.018713 12110 network_create.go:124] attempt to create docker network addons-246651 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0717 00:21:23.018762 12110 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-246651 addons-246651
I0717 00:21:23.073761 12110 network_create.go:108] docker network addons-246651 192.168.49.0/24 created
I0717 00:21:23.073804 12110 kic.go:121] calculated static IP "192.168.49.2" for the "addons-246651" container
I0717 00:21:23.073863 12110 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0717 00:21:23.087937 12110 cli_runner.go:164] Run: docker volume create addons-246651 --label name.minikube.sigs.k8s.io=addons-246651 --label created_by.minikube.sigs.k8s.io=true
I0717 00:21:23.103805 12110 oci.go:103] Successfully created a docker volume addons-246651
I0717 00:21:23.103902 12110 cli_runner.go:164] Run: docker run --rm --name addons-246651-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-246651 --entrypoint /usr/bin/test -v addons-246651:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e -d /var/lib
I0717 00:21:27.991804 12110 cli_runner.go:217] Completed: docker run --rm --name addons-246651-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-246651 --entrypoint /usr/bin/test -v addons-246651:/var gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e -d /var/lib: (4.887603479s)
I0717 00:21:27.991901 12110 oci.go:107] Successfully prepared a docker volume addons-246651
I0717 00:21:27.991936 12110 preload.go:131] Checking if preload exists for k8s version v1.30.2 and runtime docker
I0717 00:21:27.991968 12110 kic.go:194] Starting extracting preloaded images to volume ...
I0717 00:21:27.992069 12110 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19264-3553/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.30.2-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-246651:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e -I lz4 -xf /preloaded.tar -C /extractDir
I0717 00:21:32.075200 12110 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19264-3553/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.30.2-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-246651:/extractDir gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e -I lz4 -xf /preloaded.tar -C /extractDir: (4.08308848s)
I0717 00:21:32.075232 12110 kic.go:203] duration metric: took 4.083263198s to extract preloaded images to volume ...
W0717 00:21:32.075351 12110 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0717 00:21:32.075463 12110 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0717 00:21:32.121151 12110 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-246651 --name addons-246651 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-246651 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-246651 --network addons-246651 --ip 192.168.49.2 --volume addons-246651:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e
I0717 00:21:32.422134 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Running}}
I0717 00:21:32.439795 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:21:32.456655 12110 cli_runner.go:164] Run: docker exec addons-246651 stat /var/lib/dpkg/alternatives/iptables
I0717 00:21:32.496702 12110 oci.go:144] the created container "addons-246651" has a running status.
I0717 00:21:32.496736 12110 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa...
I0717 00:21:32.589614 12110 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0717 00:21:32.608677 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:21:32.624864 12110 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0717 00:21:32.624885 12110 kic_runner.go:114] Args: [docker exec --privileged addons-246651 chown docker:docker /home/docker/.ssh/authorized_keys]
I0717 00:21:32.672601 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:21:32.691717 12110 machine.go:94] provisionDockerMachine start ...
I0717 00:21:32.691818 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:32.710040 12110 main.go:141] libmachine: Using SSH client type: native
I0717 00:21:32.710332 12110 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da20] 0x830780 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0717 00:21:32.710348 12110 main.go:141] libmachine: About to run SSH command:
hostname
I0717 00:21:32.711066 12110 main.go:141] libmachine: Error dialing TCP: ssh: handshake failed: read tcp 127.0.0.1:58852->127.0.0.1:32768: read: connection reset by peer
I0717 00:21:35.841206 12110 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-246651
I0717 00:21:35.841233 12110 ubuntu.go:169] provisioning hostname "addons-246651"
I0717 00:21:35.841281 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:35.857530 12110 main.go:141] libmachine: Using SSH client type: native
I0717 00:21:35.857727 12110 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da20] 0x830780 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0717 00:21:35.857746 12110 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-246651 && echo "addons-246651" | sudo tee /etc/hostname
I0717 00:21:36.000037 12110 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-246651
I0717 00:21:36.000124 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:36.016453 12110 main.go:141] libmachine: Using SSH client type: native
I0717 00:21:36.016656 12110 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da20] 0x830780 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0717 00:21:36.016679 12110 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-246651' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-246651/g' /etc/hosts;
else
echo '127.0.1.1 addons-246651' | sudo tee -a /etc/hosts;
fi
fi
I0717 00:21:36.145625 12110 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0717 00:21:36.145655 12110 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19264-3553/.minikube CaCertPath:/home/jenkins/minikube-integration/19264-3553/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19264-3553/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19264-3553/.minikube}
I0717 00:21:36.145675 12110 ubuntu.go:177] setting up certificates
I0717 00:21:36.145688 12110 provision.go:84] configureAuth start
I0717 00:21:36.145738 12110 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-246651
I0717 00:21:36.161910 12110 provision.go:143] copyHostCerts
I0717 00:21:36.161991 12110 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19264-3553/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19264-3553/.minikube/ca.pem (1078 bytes)
I0717 00:21:36.162129 12110 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19264-3553/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19264-3553/.minikube/cert.pem (1123 bytes)
I0717 00:21:36.162206 12110 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19264-3553/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19264-3553/.minikube/key.pem (1679 bytes)
I0717 00:21:36.162268 12110 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19264-3553/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19264-3553/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19264-3553/.minikube/certs/ca-key.pem org=jenkins.addons-246651 san=[127.0.0.1 192.168.49.2 addons-246651 localhost minikube]
I0717 00:21:36.249711 12110 provision.go:177] copyRemoteCerts
I0717 00:21:36.249769 12110 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0717 00:21:36.249801 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:36.265795 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:21:36.357905 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1078 bytes)
I0717 00:21:36.377989 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0717 00:21:36.398482 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1679 bytes)
I0717 00:21:36.418599 12110 provision.go:87] duration metric: took 272.897088ms to configureAuth
I0717 00:21:36.418624 12110 ubuntu.go:193] setting minikube options for container-runtime
I0717 00:21:36.418791 12110 config.go:182] Loaded profile config "addons-246651": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.30.2
I0717 00:21:36.418834 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:36.434828 12110 main.go:141] libmachine: Using SSH client type: native
I0717 00:21:36.434994 12110 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da20] 0x830780 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0717 00:21:36.435005 12110 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0717 00:21:36.569984 12110 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0717 00:21:36.570012 12110 ubuntu.go:71] root file system type: overlay
I0717 00:21:36.570175 12110 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0717 00:21:36.570244 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:36.585855 12110 main.go:141] libmachine: Using SSH client type: native
I0717 00:21:36.586032 12110 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da20] 0x830780 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0717 00:21:36.586141 12110 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %!s(MISSING) "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0717 00:21:36.723288 12110 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0717 00:21:36.723355 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:36.739057 12110 main.go:141] libmachine: Using SSH client type: native
I0717 00:21:36.739261 12110 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82da20] 0x830780 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0717 00:21:36.739300 12110 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0717 00:21:37.403552 12110 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-06-29 00:00:53.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-07-17 00:21:36.719089813 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0717 00:21:37.403585 12110 machine.go:97] duration metric: took 4.711837453s to provisionDockerMachine
I0717 00:21:37.403598 12110 client.go:171] duration metric: took 15.147832517s to LocalClient.Create
I0717 00:21:37.403615 12110 start.go:167] duration metric: took 15.147889019s to libmachine.API.Create "addons-246651"
I0717 00:21:37.403622 12110 start.go:293] postStartSetup for "addons-246651" (driver="docker")
I0717 00:21:37.403632 12110 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0717 00:21:37.403683 12110 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0717 00:21:37.403723 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:37.419881 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:21:37.514049 12110 ssh_runner.go:195] Run: cat /etc/os-release
I0717 00:21:37.516980 12110 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0717 00:21:37.517004 12110 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0717 00:21:37.517012 12110 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0717 00:21:37.517020 12110 info.go:137] Remote host: Ubuntu 22.04.4 LTS
I0717 00:21:37.517030 12110 filesync.go:126] Scanning /home/jenkins/minikube-integration/19264-3553/.minikube/addons for local assets ...
I0717 00:21:37.517084 12110 filesync.go:126] Scanning /home/jenkins/minikube-integration/19264-3553/.minikube/files for local assets ...
I0717 00:21:37.517105 12110 start.go:296] duration metric: took 113.477253ms for postStartSetup
I0717 00:21:37.517398 12110 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-246651
I0717 00:21:37.532535 12110 profile.go:143] Saving config to /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/config.json ...
I0717 00:21:37.532777 12110 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0717 00:21:37.532824 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:37.548023 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:21:37.642325 12110 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0717 00:21:37.646112 12110 start.go:128] duration metric: took 15.392369295s to createHost
I0717 00:21:37.646134 12110 start.go:83] releasing machines lock for "addons-246651", held for 15.392514406s
I0717 00:21:37.646202 12110 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-246651
I0717 00:21:37.661433 12110 ssh_runner.go:195] Run: cat /version.json
I0717 00:21:37.661478 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:37.661511 12110 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0717 00:21:37.661592 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:21:37.677276 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:21:37.678501 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:21:37.846305 12110 ssh_runner.go:195] Run: systemctl --version
I0717 00:21:37.849958 12110 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0717 00:21:37.853670 12110 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0717 00:21:37.874276 12110 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0717 00:21:37.874340 12110 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%!p(MISSING), " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0717 00:21:37.896991 12110 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0717 00:21:37.897012 12110 start.go:495] detecting cgroup driver to use...
I0717 00:21:37.897045 12110 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0717 00:21:37.897145 12110 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0717 00:21:37.910737 12110 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.9"|' /etc/containerd/config.toml"
I0717 00:21:37.918853 12110 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0717 00:21:37.927027 12110 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0717 00:21:37.927079 12110 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0717 00:21:37.935229 12110 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0717 00:21:37.943146 12110 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0717 00:21:37.950896 12110 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0717 00:21:37.959048 12110 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0717 00:21:37.966765 12110 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0717 00:21:37.974807 12110 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0717 00:21:37.982493 12110 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0717 00:21:37.990323 12110 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0717 00:21:37.996959 12110 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0717 00:21:38.003485 12110 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0717 00:21:38.072916 12110 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0717 00:21:38.154229 12110 start.go:495] detecting cgroup driver to use...
I0717 00:21:38.154272 12110 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0717 00:21:38.154323 12110 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0717 00:21:38.164483 12110 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0717 00:21:38.164550 12110 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0717 00:21:38.174353 12110 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %!s(MISSING) "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0717 00:21:38.189054 12110 ssh_runner.go:195] Run: which cri-dockerd
I0717 00:21:38.192071 12110 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0717 00:21:38.200019 12110 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (189 bytes)
I0717 00:21:38.216722 12110 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0717 00:21:38.304710 12110 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0717 00:21:38.394162 12110 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0717 00:21:38.394295 12110 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0717 00:21:38.409784 12110 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0717 00:21:38.485165 12110 ssh_runner.go:195] Run: sudo systemctl restart docker
I0717 00:21:38.721268 12110 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0717 00:21:38.731333 12110 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0717 00:21:38.740723 12110 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0717 00:21:38.820249 12110 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0717 00:21:38.888834 12110 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0717 00:21:38.961212 12110 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0717 00:21:38.972188 12110 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0717 00:21:38.980959 12110 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0717 00:21:39.050725 12110 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0717 00:21:39.109295 12110 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0717 00:21:39.109379 12110 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0717 00:21:39.113056 12110 start.go:563] Will wait 60s for crictl version
I0717 00:21:39.113099 12110 ssh_runner.go:195] Run: which crictl
I0717 00:21:39.115989 12110 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0717 00:21:39.145553 12110 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.0.3
RuntimeApiVersion: v1
I0717 00:21:39.145611 12110 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0717 00:21:39.166181 12110 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0717 00:21:39.190238 12110 out.go:204] * Preparing Kubernetes v1.30.2 on Docker 27.0.3 ...
I0717 00:21:39.190319 12110 cli_runner.go:164] Run: docker network inspect addons-246651 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0717 00:21:39.204501 12110 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0717 00:21:39.207740 12110 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0717 00:21:39.217229 12110 kubeadm.go:883] updating cluster {Name:addons-246651 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.2 ClusterName:addons-246651 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNa
mes:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.30.2 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuF
irmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0717 00:21:39.217370 12110 preload.go:131] Checking if preload exists for k8s version v1.30.2 and runtime docker
I0717 00:21:39.217507 12110 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0717 00:21:39.233969 12110 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.30.2
registry.k8s.io/kube-controller-manager:v1.30.2
registry.k8s.io/kube-scheduler:v1.30.2
registry.k8s.io/kube-proxy:v1.30.2
registry.k8s.io/etcd:3.5.12-0
registry.k8s.io/coredns/coredns:v1.11.1
registry.k8s.io/pause:3.9
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0717 00:21:39.233992 12110 docker.go:615] Images already preloaded, skipping extraction
I0717 00:21:39.234036 12110 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0717 00:21:39.250894 12110 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.30.2
registry.k8s.io/kube-controller-manager:v1.30.2
registry.k8s.io/kube-scheduler:v1.30.2
registry.k8s.io/kube-proxy:v1.30.2
registry.k8s.io/etcd:3.5.12-0
registry.k8s.io/coredns/coredns:v1.11.1
registry.k8s.io/pause:3.9
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0717 00:21:39.250915 12110 cache_images.go:84] Images are preloaded, skipping loading
I0717 00:21:39.250934 12110 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.30.2 docker true true} ...
I0717 00:21:39.251035 12110 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.30.2/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-246651 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.30.2 ClusterName:addons-246651 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0717 00:21:39.251092 12110 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0717 00:21:39.296308 12110 cni.go:84] Creating CNI manager for ""
I0717 00:21:39.296334 12110 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0717 00:21:39.296343 12110 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0717 00:21:39.296361 12110 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.30.2 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-246651 NodeName:addons-246651 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0717 00:21:39.296490 12110 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-246651"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.30.2
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%!"(MISSING)
nodefs.inodesFree: "0%!"(MISSING)
imagefs.available: "0%!"(MISSING)
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0717 00:21:39.296557 12110 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.30.2
I0717 00:21:39.304234 12110 binaries.go:44] Found k8s binaries, skipping transfer
I0717 00:21:39.304296 12110 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0717 00:21:39.311648 12110 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0717 00:21:39.326543 12110 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0717 00:21:39.341508 12110 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0717 00:21:39.356624 12110 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0717 00:21:39.359731 12110 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0717 00:21:39.368589 12110 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0717 00:21:39.438818 12110 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0717 00:21:39.450578 12110 certs.go:68] Setting up /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651 for IP: 192.168.49.2
I0717 00:21:39.450601 12110 certs.go:194] generating shared ca certs ...
I0717 00:21:39.450628 12110 certs.go:226] acquiring lock for ca certs: {Name:mk97fac7362cad9df52e416d83559a2fa0b5f082 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:39.450759 12110 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19264-3553/.minikube/ca.key
I0717 00:21:39.525103 12110 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19264-3553/.minikube/ca.crt ...
I0717 00:21:39.525132 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/ca.crt: {Name:mk024b3995a33a36d825e6efdbebfec7e65d4e4c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:39.525306 12110 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19264-3553/.minikube/ca.key ...
I0717 00:21:39.525317 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/ca.key: {Name:mk3590babe174403b82ed88e5c6aa0873babae1f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:39.525395 12110 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19264-3553/.minikube/proxy-client-ca.key
I0717 00:21:39.692742 12110 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19264-3553/.minikube/proxy-client-ca.crt ...
I0717 00:21:39.692770 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/proxy-client-ca.crt: {Name:mk571ff0efccea1e0520c5937c365e54bd0da752 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:39.692923 12110 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19264-3553/.minikube/proxy-client-ca.key ...
I0717 00:21:39.692933 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/proxy-client-ca.key: {Name:mkf9efcf91a1b08b1b5d238d96250480cf9457b9 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:39.692997 12110 certs.go:256] generating profile certs ...
I0717 00:21:39.693046 12110 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/client.key
I0717 00:21:39.693059 12110 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/client.crt with IP's: []
I0717 00:21:39.836404 12110 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/client.crt ...
I0717 00:21:39.836430 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/client.crt: {Name:mk8fa01b24c4a5e8e76a82afdcd65590ea13443b Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:39.836575 12110 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/client.key ...
I0717 00:21:39.836588 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/client.key: {Name:mk9304ec67736c27e75495b0161f0da2b8664839 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:39.836655 12110 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.key.651fee01
I0717 00:21:39.836672 12110 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.crt.651fee01 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0717 00:21:40.083844 12110 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.crt.651fee01 ...
I0717 00:21:40.083872 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.crt.651fee01: {Name:mkb49e3ddf8840a4602a0e2141b39ffba350fec6 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:40.084018 12110 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.key.651fee01 ...
I0717 00:21:40.084031 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.key.651fee01: {Name:mkbe1402d67be459bf79a26640b0d230a43a337c Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:40.084100 12110 certs.go:381] copying /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.crt.651fee01 -> /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.crt
I0717 00:21:40.084166 12110 certs.go:385] copying /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.key.651fee01 -> /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.key
I0717 00:21:40.084210 12110 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/proxy-client.key
I0717 00:21:40.084225 12110 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/proxy-client.crt with IP's: []
I0717 00:21:40.155678 12110 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/proxy-client.crt ...
I0717 00:21:40.155702 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/proxy-client.crt: {Name:mke09aa32c00b6422b5cdd3f46a5b9a7ba189c26 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:40.155847 12110 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/proxy-client.key ...
I0717 00:21:40.155856 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/proxy-client.key: {Name:mkd511ec9aa7ad992f4f443bb0809ad1b64c793f Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:21:40.156008 12110 certs.go:484] found cert: /home/jenkins/minikube-integration/19264-3553/.minikube/certs/ca-key.pem (1679 bytes)
I0717 00:21:40.156038 12110 certs.go:484] found cert: /home/jenkins/minikube-integration/19264-3553/.minikube/certs/ca.pem (1078 bytes)
I0717 00:21:40.156062 12110 certs.go:484] found cert: /home/jenkins/minikube-integration/19264-3553/.minikube/certs/cert.pem (1123 bytes)
I0717 00:21:40.156086 12110 certs.go:484] found cert: /home/jenkins/minikube-integration/19264-3553/.minikube/certs/key.pem (1679 bytes)
I0717 00:21:40.156616 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0717 00:21:40.178031 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0717 00:21:40.197869 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0717 00:21:40.217710 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1679 bytes)
I0717 00:21:40.237306 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0717 00:21:40.256570 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1679 bytes)
I0717 00:21:40.275778 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0717 00:21:40.295687 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/profiles/addons-246651/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I0717 00:21:40.315039 12110 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19264-3553/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0717 00:21:40.334802 12110 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0717 00:21:40.350445 12110 ssh_runner.go:195] Run: openssl version
I0717 00:21:40.355792 12110 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0717 00:21:40.364356 12110 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0717 00:21:40.367606 12110 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Jul 17 00:21 /usr/share/ca-certificates/minikubeCA.pem
I0717 00:21:40.367660 12110 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0717 00:21:40.373662 12110 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0717 00:21:40.381458 12110 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0717 00:21:40.384158 12110 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0717 00:21:40.384213 12110 kubeadm.go:392] StartCluster: {Name:addons-246651 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase-builds:v0.0.44-1721146479-19264@sha256:7ee06b7e8fb4a6c7fce11a567253ea7d43fed61ee0beca281a1ac2c2566a2a2e Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.30.2 ClusterName:addons-246651 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames
:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.30.2 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirm
warePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0717 00:21:40.384316 12110 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0717 00:21:40.401627 12110 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0717 00:21:40.409133 12110 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0717 00:21:40.416730 12110 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0717 00:21:40.416771 12110 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0717 00:21:40.423737 12110 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0717 00:21:40.423754 12110 kubeadm.go:157] found existing configuration files:
I0717 00:21:40.423803 12110 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0717 00:21:40.431124 12110 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0717 00:21:40.431176 12110 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0717 00:21:40.438172 12110 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0717 00:21:40.445064 12110 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0717 00:21:40.445120 12110 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0717 00:21:40.451877 12110 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0717 00:21:40.458922 12110 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0717 00:21:40.458974 12110 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0717 00:21:40.466047 12110 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0717 00:21:40.473095 12110 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0717 00:21:40.473145 12110 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0717 00:21:40.479931 12110 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.30.2:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0717 00:21:40.549631 12110 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1062-gcp\n", err: exit status 1
I0717 00:21:40.600491 12110 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0717 00:21:49.626410 12110 kubeadm.go:310] [init] Using Kubernetes version: v1.30.2
I0717 00:21:49.626536 12110 kubeadm.go:310] [preflight] Running pre-flight checks
I0717 00:21:49.626699 12110 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0717 00:21:49.626760 12110 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1062-gcp[0m
I0717 00:21:49.626791 12110 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0717 00:21:49.626830 12110 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0717 00:21:49.626871 12110 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0717 00:21:49.626911 12110 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0717 00:21:49.626992 12110 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0717 00:21:49.627066 12110 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0717 00:21:49.627115 12110 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0717 00:21:49.627154 12110 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0717 00:21:49.627202 12110 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0717 00:21:49.627256 12110 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0717 00:21:49.627383 12110 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0717 00:21:49.627524 12110 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0717 00:21:49.627636 12110 kubeadm.go:310] [preflight] You can also perform this action in beforehand using 'kubeadm config images pull'
I0717 00:21:49.627695 12110 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0717 00:21:49.629023 12110 out.go:204] - Generating certificates and keys ...
I0717 00:21:49.629088 12110 kubeadm.go:310] [certs] Using existing ca certificate authority
I0717 00:21:49.629151 12110 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0717 00:21:49.629212 12110 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0717 00:21:49.629260 12110 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0717 00:21:49.629323 12110 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0717 00:21:49.629368 12110 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0717 00:21:49.629432 12110 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0717 00:21:49.629562 12110 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-246651 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0717 00:21:49.629631 12110 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0717 00:21:49.629771 12110 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-246651 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0717 00:21:49.629839 12110 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0717 00:21:49.629919 12110 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0717 00:21:49.629986 12110 kubeadm.go:310] [certs] Generating "sa" key and public key
I0717 00:21:49.630077 12110 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0717 00:21:49.630144 12110 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0717 00:21:49.630220 12110 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0717 00:21:49.630269 12110 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0717 00:21:49.630353 12110 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0717 00:21:49.630432 12110 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0717 00:21:49.630543 12110 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0717 00:21:49.630625 12110 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0717 00:21:49.632720 12110 out.go:204] - Booting up control plane ...
I0717 00:21:49.632808 12110 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0717 00:21:49.632923 12110 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0717 00:21:49.633001 12110 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0717 00:21:49.633144 12110 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0717 00:21:49.633216 12110 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0717 00:21:49.633249 12110 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0717 00:21:49.633369 12110 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0717 00:21:49.633440 12110 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet. This can take up to 4m0s
I0717 00:21:49.633495 12110 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 501.948299ms
I0717 00:21:49.633552 12110 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0717 00:21:49.633608 12110 kubeadm.go:310] [api-check] The API server is healthy after 4.501893857s
I0717 00:21:49.633693 12110 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0717 00:21:49.633797 12110 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0717 00:21:49.633848 12110 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0717 00:21:49.634003 12110 kubeadm.go:310] [mark-control-plane] Marking the node addons-246651 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0717 00:21:49.634050 12110 kubeadm.go:310] [bootstrap-token] Using token: tr0pr4.wyyg1btlug0psz0c
I0717 00:21:49.635198 12110 out.go:204] - Configuring RBAC rules ...
I0717 00:21:49.635297 12110 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0717 00:21:49.635387 12110 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0717 00:21:49.635548 12110 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0717 00:21:49.635724 12110 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0717 00:21:49.635886 12110 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0717 00:21:49.636002 12110 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0717 00:21:49.636098 12110 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0717 00:21:49.636139 12110 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0717 00:21:49.636194 12110 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0717 00:21:49.636207 12110 kubeadm.go:310]
I0717 00:21:49.636303 12110 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0717 00:21:49.636314 12110 kubeadm.go:310]
I0717 00:21:49.636410 12110 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0717 00:21:49.636420 12110 kubeadm.go:310]
I0717 00:21:49.636441 12110 kubeadm.go:310] mkdir -p $HOME/.kube
I0717 00:21:49.636491 12110 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0717 00:21:49.636539 12110 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0717 00:21:49.636551 12110 kubeadm.go:310]
I0717 00:21:49.636629 12110 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0717 00:21:49.636639 12110 kubeadm.go:310]
I0717 00:21:49.636705 12110 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0717 00:21:49.636713 12110 kubeadm.go:310]
I0717 00:21:49.636791 12110 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0717 00:21:49.636857 12110 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0717 00:21:49.636946 12110 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0717 00:21:49.636960 12110 kubeadm.go:310]
I0717 00:21:49.637070 12110 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0717 00:21:49.637170 12110 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0717 00:21:49.637183 12110 kubeadm.go:310]
I0717 00:21:49.637285 12110 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token tr0pr4.wyyg1btlug0psz0c \
I0717 00:21:49.637411 12110 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:a90a59c971912bbf3fd6ec54ae4a512153369ffdc9dcccebbf228da9d3c5ffad \
I0717 00:21:49.637442 12110 kubeadm.go:310] --control-plane
I0717 00:21:49.637451 12110 kubeadm.go:310]
I0717 00:21:49.637550 12110 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0717 00:21:49.637557 12110 kubeadm.go:310]
I0717 00:21:49.637657 12110 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token tr0pr4.wyyg1btlug0psz0c \
I0717 00:21:49.637801 12110 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:a90a59c971912bbf3fd6ec54ae4a512153369ffdc9dcccebbf228da9d3c5ffad
I0717 00:21:49.637818 12110 cni.go:84] Creating CNI manager for ""
I0717 00:21:49.637832 12110 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0717 00:21:49.639104 12110 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0717 00:21:49.640196 12110 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0717 00:21:49.647999 12110 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0717 00:21:49.663711 12110 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0717 00:21:49.663775 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:49.663786 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-246651 minikube.k8s.io/updated_at=2024_07_17T00_21_49_0700 minikube.k8s.io/version=v1.33.1 minikube.k8s.io/commit=3cfbbb17fd76400a5ee2ea427db7148a0ef7c185 minikube.k8s.io/name=addons-246651 minikube.k8s.io/primary=true
I0717 00:21:49.670465 12110 ops.go:34] apiserver oom_adj: -16
I0717 00:21:49.736254 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:50.236753 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:50.736641 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:51.237170 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:51.736576 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:52.236498 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:52.736968 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:53.236358 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:53.737052 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:54.236492 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:54.737008 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:55.237237 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:55.736981 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:56.236422 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:56.737358 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:57.236692 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:57.736791 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:58.236763 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:58.736443 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:59.236377 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:21:59.736283 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:22:00.236505 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:22:00.737177 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:22:01.237013 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:22:01.736292 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:22:02.236817 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:22:02.736816 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:22:03.237147 12110 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.30.2/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0717 00:22:03.485627 12110 kubeadm.go:1113] duration metric: took 13.821903661s to wait for elevateKubeSystemPrivileges
I0717 00:22:03.485712 12110 kubeadm.go:394] duration metric: took 23.101501563s to StartCluster
I0717 00:22:03.485739 12110 settings.go:142] acquiring lock: {Name:mkb057382da6577dbb2fc944ce0cc77f8d4e599e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:22:03.485856 12110 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19264-3553/kubeconfig
I0717 00:22:03.486372 12110 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19264-3553/kubeconfig: {Name:mka9a45b08e1b98005a35fab6350dd0543e49a00 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0717 00:22:03.486645 12110 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.30.2 ContainerRuntime:docker ControlPlane:true Worker:true}
I0717 00:22:03.486730 12110 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.30.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0717 00:22:03.486736 12110 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0717 00:22:03.486825 12110 addons.go:69] Setting yakd=true in profile "addons-246651"
I0717 00:22:03.486840 12110 addons.go:69] Setting inspektor-gadget=true in profile "addons-246651"
I0717 00:22:03.486854 12110 addons.go:234] Setting addon yakd=true in "addons-246651"
I0717 00:22:03.486864 12110 addons.go:234] Setting addon inspektor-gadget=true in "addons-246651"
I0717 00:22:03.486886 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.486895 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.486908 12110 config.go:182] Loaded profile config "addons-246651": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.30.2
I0717 00:22:03.486906 12110 addons.go:69] Setting storage-provisioner=true in profile "addons-246651"
I0717 00:22:03.486942 12110 addons.go:69] Setting gcp-auth=true in profile "addons-246651"
I0717 00:22:03.486954 12110 addons.go:234] Setting addon storage-provisioner=true in "addons-246651"
I0717 00:22:03.486959 12110 mustload.go:65] Loading cluster: addons-246651
I0717 00:22:03.486990 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.487006 12110 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-246651"
I0717 00:22:03.487028 12110 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-246651"
I0717 00:22:03.487114 12110 config.go:182] Loaded profile config "addons-246651": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.30.2
I0717 00:22:03.487144 12110 addons.go:69] Setting ingress=true in profile "addons-246651"
I0717 00:22:03.487191 12110 addons.go:234] Setting addon ingress=true in "addons-246651"
I0717 00:22:03.487223 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.487271 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.487344 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.487398 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.487469 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.487606 12110 addons.go:69] Setting helm-tiller=true in profile "addons-246651"
I0717 00:22:03.487634 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.487667 12110 addons.go:234] Setting addon helm-tiller=true in "addons-246651"
I0717 00:22:03.487711 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.487979 12110 addons.go:69] Setting cloud-spanner=true in profile "addons-246651"
I0717 00:22:03.488014 12110 addons.go:234] Setting addon cloud-spanner=true in "addons-246651"
I0717 00:22:03.488048 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.488205 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.488219 12110 addons.go:69] Setting ingress-dns=true in profile "addons-246651"
I0717 00:22:03.488241 12110 addons.go:234] Setting addon ingress-dns=true in "addons-246651"
I0717 00:22:03.488268 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.488289 12110 addons.go:69] Setting volcano=true in profile "addons-246651"
I0717 00:22:03.488328 12110 addons.go:234] Setting addon volcano=true in "addons-246651"
I0717 00:22:03.488357 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.488450 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.488653 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.488788 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.494122 12110 out.go:177] * Verifying Kubernetes components...
I0717 00:22:03.499268 12110 addons.go:69] Setting volumesnapshots=true in profile "addons-246651"
I0717 00:22:03.499381 12110 addons.go:234] Setting addon volumesnapshots=true in "addons-246651"
I0717 00:22:03.499455 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.499738 12110 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-246651"
I0717 00:22:03.500055 12110 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-246651"
I0717 00:22:03.500132 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.500749 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.500091 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.501416 12110 addons.go:69] Setting default-storageclass=true in profile "addons-246651"
I0717 00:22:03.501656 12110 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-246651"
I0717 00:22:03.502016 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.488213 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.505378 12110 addons.go:69] Setting metrics-server=true in profile "addons-246651"
I0717 00:22:03.505423 12110 addons.go:234] Setting addon metrics-server=true in "addons-246651"
I0717 00:22:03.505471 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.505607 12110 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-246651"
I0717 00:22:03.505670 12110 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-246651"
I0717 00:22:03.505730 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.506014 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.505674 12110 addons.go:69] Setting registry=true in profile "addons-246651"
I0717 00:22:03.506270 12110 addons.go:234] Setting addon registry=true in "addons-246651"
I0717 00:22:03.506304 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.506344 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.501558 12110 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0717 00:22:03.533589 12110 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-246651"
I0717 00:22:03.533648 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.534964 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.536311 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.538365 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.540975 12110 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.30.0
I0717 00:22:03.546136 12110 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0717 00:22:03.546158 12110 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0717 00:22:03.546234 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.552365 12110 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0717 00:22:03.555094 12110 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.17
I0717 00:22:03.555171 12110 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0717 00:22:03.555207 12110 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0717 00:22:03.556327 12110 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0717 00:22:03.556530 12110 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0717 00:22:03.556561 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0717 00:22:03.556532 12110 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0717 00:22:03.556608 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0717 00:22:03.556619 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.556675 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.556780 12110 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0717 00:22:03.556789 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0717 00:22:03.556828 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.557834 12110 addons.go:234] Setting addon default-storageclass=true in "addons-246651"
I0717 00:22:03.557875 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:03.558328 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:03.559447 12110 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0717 00:22:03.564102 12110 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0717 00:22:03.564128 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0717 00:22:03.564188 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.571418 12110 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0717 00:22:03.572386 12110 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0717 00:22:03.573292 12110 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0717 00:22:03.573311 12110 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0717 00:22:03.573370 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.573409 12110 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0717 00:22:03.573420 12110 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0717 00:22:03.573461 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.576930 12110 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0717 00:22:03.578285 12110 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.10.1
I0717 00:22:03.579400 12110 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.1
I0717 00:22:03.580415 12110 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.1
I0717 00:22:03.580470 12110 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0717 00:22:03.581593 12110 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0717 00:22:03.581614 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0717 00:22:03.581675 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.584822 12110 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0717 00:22:03.585973 12110 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0717 00:22:03.587453 12110 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0717 00:22:03.588694 12110 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0717 00:22:03.588779 12110 out.go:177] - Using image docker.io/busybox:stable
I0717 00:22:03.589763 12110 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0717 00:22:03.589779 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0717 00:22:03.589834 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.590009 12110 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0717 00:22:03.591076 12110 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0717 00:22:03.592166 12110 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0717 00:22:03.593255 12110 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0717 00:22:03.593271 12110 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0717 00:22:03.593334 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.593624 12110 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0717 00:22:03.594560 12110 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.0
I0717 00:22:03.595654 12110 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0717 00:22:03.595676 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0717 00:22:03.595730 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.598228 12110 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0717 00:22:03.598244 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0717 00:22:03.598292 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.615354 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.621811 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.630138 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.639797 12110 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0717 00:22:03.641065 12110 out.go:177] - Using image docker.io/registry:2.8.3
I0717 00:22:03.642258 12110 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0717 00:22:03.642277 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (798 bytes)
I0717 00:22:03.642336 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.645395 12110 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0717 00:22:03.645410 12110 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0717 00:22:03.645467 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.646650 12110 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.1
I0717 00:22:03.647950 12110 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0717 00:22:03.647969 12110 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0717 00:22:03.648022 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:03.649430 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.651610 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.654288 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.654716 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.662671 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.663051 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.663092 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.664118 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.664587 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.668884 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.672652 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:03.676406 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
W0717 00:22:03.694389 12110 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0717 00:22:03.694428 12110 retry.go:31] will retry after 218.976652ms: ssh: handshake failed: EOF
W0717 00:22:03.694393 12110 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0717 00:22:03.694446 12110 retry.go:31] will retry after 330.799139ms: ssh: handshake failed: EOF
W0717 00:22:03.984663 12110 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0717 00:22:03.984701 12110 retry.go:31] will retry after 312.54509ms: ssh: handshake failed: EOF
I0717 00:22:04.093933 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0717 00:22:04.099563 12110 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0717 00:22:04.099597 12110 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.30.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.30.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0717 00:22:04.198193 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0717 00:22:04.199816 12110 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0717 00:22:04.199879 12110 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0717 00:22:04.285441 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0717 00:22:04.292893 12110 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0717 00:22:04.292986 12110 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0717 00:22:04.301243 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0717 00:22:04.301725 12110 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0717 00:22:04.301785 12110 ssh_runner.go:362] scp helm-tiller/helm-tiller-rbac.yaml --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0717 00:22:04.302706 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0717 00:22:04.383584 12110 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0717 00:22:04.383678 12110 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0717 00:22:04.394094 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0717 00:22:04.397199 12110 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0717 00:22:04.397280 12110 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0717 00:22:04.397905 12110 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0717 00:22:04.397947 12110 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0717 00:22:04.404779 12110 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0717 00:22:04.404806 12110 ssh_runner.go:362] scp helm-tiller/helm-tiller-svc.yaml --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0717 00:22:04.405467 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0717 00:22:04.491107 12110 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0717 00:22:04.491138 12110 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0717 00:22:04.587726 12110 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0717 00:22:04.587769 12110 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0717 00:22:04.592736 12110 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0717 00:22:04.592806 12110 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0717 00:22:04.684194 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0717 00:22:04.694269 12110 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0717 00:22:04.694351 12110 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0717 00:22:04.790789 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0717 00:22:04.791388 12110 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0717 00:22:04.791451 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0717 00:22:04.983869 12110 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0717 00:22:04.983900 12110 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0717 00:22:04.996806 12110 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0717 00:22:04.996832 12110 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0717 00:22:04.997687 12110 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0717 00:22:04.997700 12110 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0717 00:22:05.004084 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0717 00:22:05.096659 12110 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0717 00:22:05.096689 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0717 00:22:05.194904 12110 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0717 00:22:05.194935 12110 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0717 00:22:05.300535 12110 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0717 00:22:05.300576 12110 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0717 00:22:05.399987 12110 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0717 00:22:05.400015 12110 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0717 00:22:05.598653 12110 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0717 00:22:05.598682 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0717 00:22:05.602655 12110 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0717 00:22:05.602679 12110 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0717 00:22:05.691130 12110 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0717 00:22:05.691185 12110 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0717 00:22:05.789460 12110 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0717 00:22:05.789488 12110 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0717 00:22:05.796847 12110 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0717 00:22:05.796877 12110 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0717 00:22:05.802843 12110 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0717 00:22:05.802872 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0717 00:22:05.983184 12110 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0717 00:22:05.983226 12110 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0717 00:22:05.986425 12110 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0717 00:22:05.986450 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0717 00:22:05.986994 12110 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0717 00:22:05.987025 12110 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0717 00:22:06.187737 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0717 00:22:06.197507 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0717 00:22:06.285766 12110 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0717 00:22:06.285800 12110 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0717 00:22:06.400574 12110 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0717 00:22:06.400606 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0717 00:22:06.488955 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0717 00:22:06.491580 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (2.397556799s)
I0717 00:22:06.491662 12110 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.30.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.30.2/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (2.391927762s)
I0717 00:22:06.491683 12110 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0717 00:22:06.492897 12110 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (2.39324819s)
I0717 00:22:06.493806 12110 node_ready.go:35] waiting up to 6m0s for node "addons-246651" to be "Ready" ...
I0717 00:22:06.496308 12110 node_ready.go:49] node "addons-246651" has status "Ready":"True"
I0717 00:22:06.496339 12110 node_ready.go:38] duration metric: took 2.505235ms for node "addons-246651" to be "Ready" ...
I0717 00:22:06.496350 12110 pod_ready.go:35] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0717 00:22:06.601130 12110 pod_ready.go:78] waiting up to 6m0s for pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace to be "Ready" ...
I0717 00:22:06.701358 12110 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0717 00:22:06.701389 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0717 00:22:06.990109 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0717 00:22:06.997719 12110 kapi.go:248] "coredns" deployment in "kube-system" namespace and "addons-246651" context rescaled to 1 replicas
I0717 00:22:07.103752 12110 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0717 00:22:07.103779 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0717 00:22:07.285927 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (3.087642131s)
I0717 00:22:07.489003 12110 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0717 00:22:07.489028 12110 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0717 00:22:07.603958 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (3.302603302s)
I0717 00:22:07.604034 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (3.318252259s)
I0717 00:22:08.085786 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0717 00:22:08.607078 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:10.590926 12110 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0717 00:22:10.591056 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:10.618376 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:10.688517 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:11.205733 12110 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0717 00:22:11.307808 12110 addons.go:234] Setting addon gcp-auth=true in "addons-246651"
I0717 00:22:11.307856 12110 host.go:66] Checking if "addons-246651" exists ...
I0717 00:22:11.308223 12110 cli_runner.go:164] Run: docker container inspect addons-246651 --format={{.State.Status}}
I0717 00:22:11.324836 12110 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0717 00:22:11.324890 12110 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-246651
I0717 00:22:11.340345 12110 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19264-3553/.minikube/machines/addons-246651/id_rsa Username:docker}
I0717 00:22:12.387215 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (7.993026719s)
I0717 00:22:12.387342 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (7.981845689s)
I0717 00:22:12.387421 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (7.703116465s)
I0717 00:22:12.387559 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (8.084796887s)
I0717 00:22:12.387586 12110 addons.go:475] Verifying addon ingress=true in "addons-246651"
I0717 00:22:12.389497 12110 out.go:177] * Verifying ingress addon...
I0717 00:22:12.391730 12110 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
I0717 00:22:12.398610 12110 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0717 00:22:12.398633 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:12.901785 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:13.183999 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:13.396816 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:13.990501 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:14.486396 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:14.896991 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:15.185344 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:15.298253 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (10.507405514s)
I0717 00:22:15.298317 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (10.294194503s)
I0717 00:22:15.298380 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (9.110611007s)
I0717 00:22:15.298391 12110 addons.go:475] Verifying addon registry=true in "addons-246651"
I0717 00:22:15.298523 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (9.100981381s)
W0717 00:22:15.298558 12110 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0717 00:22:15.298606 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (8.809614564s)
I0717 00:22:15.298624 12110 addons.go:475] Verifying addon metrics-server=true in "addons-246651"
I0717 00:22:15.298578 12110 retry.go:31] will retry after 202.545027ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0717 00:22:15.298709 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (8.308562193s)
I0717 00:22:15.301201 12110 out.go:177] * Verifying registry addon...
I0717 00:22:15.301297 12110 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-246651 service yakd-dashboard -n yakd-dashboard
I0717 00:22:15.305803 12110 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0717 00:22:15.387511 12110 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0717 00:22:15.387595 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:15.395795 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:15.501875 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0717 00:22:15.885978 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:15.901069 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:16.385061 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:16.398039 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:16.582684 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (8.496845656s)
I0717 00:22:16.582763 12110 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-246651"
I0717 00:22:16.582726 12110 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.257865862s)
I0717 00:22:16.584529 12110 out.go:177] * Verifying csi-hostpath-driver addon...
I0717 00:22:16.584529 12110 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.1
I0717 00:22:16.586094 12110 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0717 00:22:16.586967 12110 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0717 00:22:16.587645 12110 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0717 00:22:16.587666 12110 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0717 00:22:16.593432 12110 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0717 00:22:16.593498 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:16.693717 12110 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0717 00:22:16.693747 12110 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0717 00:22:16.784989 12110 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0717 00:22:16.785015 12110 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0717 00:22:16.808486 12110 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0717 00:22:16.811092 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:16.897135 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:17.093737 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:17.384155 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:17.396696 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:17.593418 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:17.607205 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:17.810475 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:17.896424 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:17.992924 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.490992585s)
I0717 00:22:18.093253 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:18.207493 12110 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.30.2/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.398964084s)
I0717 00:22:18.209492 12110 addons.go:475] Verifying addon gcp-auth=true in "addons-246651"
I0717 00:22:18.211121 12110 out.go:177] * Verifying gcp-auth addon...
I0717 00:22:18.213164 12110 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0717 00:22:18.215003 12110 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0717 00:22:18.310424 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:18.395523 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:18.593473 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:18.811281 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:18.895659 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:19.091813 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:19.310692 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:19.395924 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:19.592239 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:19.809925 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:19.896457 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:20.094217 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:20.183889 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:20.310796 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:20.396332 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:20.592547 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:20.810853 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:20.896278 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:21.093033 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:21.310684 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:21.396051 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:21.591626 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:21.912452 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:21.912749 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:22.091781 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:22.310288 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:22.395679 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:22.592443 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:22.606299 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:22.809444 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:22.896480 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:23.091694 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:23.310205 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:23.395261 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:23.592028 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:23.811603 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:23.895526 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:24.092296 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:24.310156 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:24.396067 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:24.592928 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:24.606720 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:24.810461 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:24.896175 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:25.092920 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:25.310931 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:25.396391 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:25.592751 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:25.810484 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:25.895598 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:26.092556 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:26.310727 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:26.395846 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:26.592128 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:26.606918 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:26.810317 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:26.895823 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:27.093125 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:27.310329 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:27.395506 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:27.592861 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:27.811386 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:27.895451 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:28.097997 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:28.309868 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:28.396034 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:28.592428 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:28.809781 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:28.896446 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:29.092373 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:29.106002 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:29.309906 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:29.396130 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:29.591788 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:29.810872 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:29.895461 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:30.091361 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:30.309291 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:30.396032 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:30.593700 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:30.810714 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:30.896711 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:31.092153 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:31.309490 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:31.395911 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:31.591478 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:31.606150 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:31.810535 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:31.895726 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:32.092323 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:32.309899 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:32.396618 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:32.591848 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:32.885453 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:32.895774 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:33.093681 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:33.310946 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:33.396478 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:33.596530 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:33.809670 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:33.895691 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:34.091473 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:34.105696 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:34.310345 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:34.395454 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:34.592419 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:34.809739 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:34.896251 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:35.092908 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:35.310079 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:35.395732 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:35.591495 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:35.810145 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:35.895991 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:36.092289 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:36.106723 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:36.311305 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:36.396999 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:36.593108 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:36.810533 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:36.896161 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:37.091725 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:37.311101 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:37.396162 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:37.591933 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:37.809420 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:37.895554 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:38.091510 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:38.309663 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:38.395678 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:38.592471 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:38.606181 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:38.810328 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:38.896911 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:39.091634 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:39.309578 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:39.395809 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:39.591470 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:39.810183 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:39.895181 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:40.091509 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:40.310057 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:40.396090 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:40.592028 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:40.606921 12110 pod_ready.go:102] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"False"
I0717 00:22:40.809403 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:40.895633 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:41.092691 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:41.310693 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:41.396041 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:41.592551 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:41.810231 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:41.895851 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:42.091827 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:42.309661 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:42.396113 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:42.592562 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:42.811161 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:42.900241 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:43.092139 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:43.106769 12110 pod_ready.go:92] pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace has status "Ready":"True"
I0717 00:22:43.106794 12110 pod_ready.go:81] duration metric: took 36.505627781s for pod "coredns-7db6d8ff4d-2hm4d" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.106806 12110 pod_ready.go:78] waiting up to 6m0s for pod "coredns-7db6d8ff4d-j8kgf" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.108516 12110 pod_ready.go:97] error getting pod "coredns-7db6d8ff4d-j8kgf" in "kube-system" namespace (skipping!): pods "coredns-7db6d8ff4d-j8kgf" not found
I0717 00:22:43.108540 12110 pod_ready.go:81] duration metric: took 1.726845ms for pod "coredns-7db6d8ff4d-j8kgf" in "kube-system" namespace to be "Ready" ...
E0717 00:22:43.108552 12110 pod_ready.go:66] WaitExtra: waitPodCondition: error getting pod "coredns-7db6d8ff4d-j8kgf" in "kube-system" namespace (skipping!): pods "coredns-7db6d8ff4d-j8kgf" not found
I0717 00:22:43.108561 12110 pod_ready.go:78] waiting up to 6m0s for pod "etcd-addons-246651" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.112524 12110 pod_ready.go:92] pod "etcd-addons-246651" in "kube-system" namespace has status "Ready":"True"
I0717 00:22:43.112544 12110 pod_ready.go:81] duration metric: took 3.974768ms for pod "etcd-addons-246651" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.112555 12110 pod_ready.go:78] waiting up to 6m0s for pod "kube-apiserver-addons-246651" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.116574 12110 pod_ready.go:92] pod "kube-apiserver-addons-246651" in "kube-system" namespace has status "Ready":"True"
I0717 00:22:43.116596 12110 pod_ready.go:81] duration metric: took 4.028022ms for pod "kube-apiserver-addons-246651" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.116607 12110 pod_ready.go:78] waiting up to 6m0s for pod "kube-controller-manager-addons-246651" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.120552 12110 pod_ready.go:92] pod "kube-controller-manager-addons-246651" in "kube-system" namespace has status "Ready":"True"
I0717 00:22:43.120573 12110 pod_ready.go:81] duration metric: took 3.957411ms for pod "kube-controller-manager-addons-246651" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.120584 12110 pod_ready.go:78] waiting up to 6m0s for pod "kube-proxy-nzcqj" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.305526 12110 pod_ready.go:92] pod "kube-proxy-nzcqj" in "kube-system" namespace has status "Ready":"True"
I0717 00:22:43.305553 12110 pod_ready.go:81] duration metric: took 184.961005ms for pod "kube-proxy-nzcqj" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.305566 12110 pod_ready.go:78] waiting up to 6m0s for pod "kube-scheduler-addons-246651" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.310109 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:43.396130 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:43.591842 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:43.704500 12110 pod_ready.go:92] pod "kube-scheduler-addons-246651" in "kube-system" namespace has status "Ready":"True"
I0717 00:22:43.704522 12110 pod_ready.go:81] duration metric: took 398.947516ms for pod "kube-scheduler-addons-246651" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.704532 12110 pod_ready.go:78] waiting up to 6m0s for pod "nvidia-device-plugin-daemonset-g6hlf" in "kube-system" namespace to be "Ready" ...
I0717 00:22:43.809985 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:43.896009 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:44.092472 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:44.104913 12110 pod_ready.go:92] pod "nvidia-device-plugin-daemonset-g6hlf" in "kube-system" namespace has status "Ready":"True"
I0717 00:22:44.104934 12110 pod_ready.go:81] duration metric: took 400.395064ms for pod "nvidia-device-plugin-daemonset-g6hlf" in "kube-system" namespace to be "Ready" ...
I0717 00:22:44.104943 12110 pod_ready.go:38] duration metric: took 37.608577999s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0717 00:22:44.104964 12110 api_server.go:52] waiting for apiserver process to appear ...
I0717 00:22:44.105023 12110 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0717 00:22:44.119358 12110 api_server.go:72] duration metric: took 40.632671282s to wait for apiserver process to appear ...
I0717 00:22:44.119386 12110 api_server.go:88] waiting for apiserver healthz status ...
I0717 00:22:44.119411 12110 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0717 00:22:44.122743 12110 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0717 00:22:44.123550 12110 api_server.go:141] control plane version: v1.30.2
I0717 00:22:44.123574 12110 api_server.go:131] duration metric: took 4.180539ms to wait for apiserver health ...
I0717 00:22:44.123584 12110 system_pods.go:43] waiting for kube-system pods to appear ...
I0717 00:22:44.311621 12110 system_pods.go:59] 18 kube-system pods found
I0717 00:22:44.311648 12110 system_pods.go:61] "coredns-7db6d8ff4d-2hm4d" [6d9c5694-2238-4d9b-a8cc-bad389b1a061] Running
I0717 00:22:44.311656 12110 system_pods.go:61] "csi-hostpath-attacher-0" [49c4c1f5-514d-4e0b-8d5c-18e8afdc9cb9] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0717 00:22:44.311663 12110 system_pods.go:61] "csi-hostpath-resizer-0" [5996a5ca-3dbd-49c9-a61f-51af4bdce479] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0717 00:22:44.311670 12110 system_pods.go:61] "csi-hostpathplugin-pn6m2" [a74e73be-eded-4659-8ba4-a8038ecf32e3] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0717 00:22:44.311674 12110 system_pods.go:61] "etcd-addons-246651" [6a277c3e-84c1-4bec-bff3-875e36ba54c3] Running
I0717 00:22:44.311678 12110 system_pods.go:61] "kube-apiserver-addons-246651" [2a8a29c3-67ba-40f5-b986-68c4da1729a9] Running
I0717 00:22:44.311681 12110 system_pods.go:61] "kube-controller-manager-addons-246651" [628c8d4b-3ef2-4ac9-9594-d919b039c856] Running
I0717 00:22:44.311685 12110 system_pods.go:61] "kube-ingress-dns-minikube" [0ed9119c-c7eb-4c2f-ba3c-79fea7e2ba91] Running
I0717 00:22:44.311689 12110 system_pods.go:61] "kube-proxy-nzcqj" [0d972af8-6f49-43d1-84ed-0e938d71043b] Running
I0717 00:22:44.311693 12110 system_pods.go:61] "kube-scheduler-addons-246651" [d6598e71-3c71-43b0-b05d-5f199d4e0c34] Running
I0717 00:22:44.311697 12110 system_pods.go:61] "metrics-server-c59844bb4-2vpkj" [e3e50235-5bcc-435e-b32a-3db12ef1208a] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0717 00:22:44.311704 12110 system_pods.go:61] "nvidia-device-plugin-daemonset-g6hlf" [8c31d1b6-e782-4497-b523-0002031aa0c8] Running
I0717 00:22:44.311707 12110 system_pods.go:61] "registry-mwnw8" [922f7c1c-219a-49d4-9764-6789d827e7e8] Running
I0717 00:22:44.311712 12110 system_pods.go:61] "registry-proxy-crj68" [23346467-00bb-4945-bb70-82048f3eb257] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0717 00:22:44.311719 12110 system_pods.go:61] "snapshot-controller-745499f584-x95mr" [71c7a587-246c-4445-b115-08a1d678211d] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0717 00:22:44.311727 12110 system_pods.go:61] "snapshot-controller-745499f584-xmg7r" [164aa91a-825f-4cbc-a1cf-96f7d7fc5391] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0717 00:22:44.311731 12110 system_pods.go:61] "storage-provisioner" [1518d476-d81f-4a67-aed7-5f0ead3ab81c] Running
I0717 00:22:44.311734 12110 system_pods.go:61] "tiller-deploy-6677d64bcd-6llsw" [d8908f47-fb82-48b4-881d-6bd342a17f86] Running
I0717 00:22:44.311740 12110 system_pods.go:74] duration metric: took 188.150767ms to wait for pod list to return data ...
I0717 00:22:44.311749 12110 default_sa.go:34] waiting for default service account to be created ...
I0717 00:22:44.312318 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:44.395609 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:44.504690 12110 default_sa.go:45] found service account: "default"
I0717 00:22:44.504719 12110 default_sa.go:55] duration metric: took 192.960856ms for default service account to be created ...
I0717 00:22:44.504730 12110 system_pods.go:116] waiting for k8s-apps to be running ...
I0717 00:22:44.591964 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:44.711130 12110 system_pods.go:86] 18 kube-system pods found
I0717 00:22:44.711156 12110 system_pods.go:89] "coredns-7db6d8ff4d-2hm4d" [6d9c5694-2238-4d9b-a8cc-bad389b1a061] Running
I0717 00:22:44.711165 12110 system_pods.go:89] "csi-hostpath-attacher-0" [49c4c1f5-514d-4e0b-8d5c-18e8afdc9cb9] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0717 00:22:44.711171 12110 system_pods.go:89] "csi-hostpath-resizer-0" [5996a5ca-3dbd-49c9-a61f-51af4bdce479] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0717 00:22:44.711179 12110 system_pods.go:89] "csi-hostpathplugin-pn6m2" [a74e73be-eded-4659-8ba4-a8038ecf32e3] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0717 00:22:44.711183 12110 system_pods.go:89] "etcd-addons-246651" [6a277c3e-84c1-4bec-bff3-875e36ba54c3] Running
I0717 00:22:44.711188 12110 system_pods.go:89] "kube-apiserver-addons-246651" [2a8a29c3-67ba-40f5-b986-68c4da1729a9] Running
I0717 00:22:44.711191 12110 system_pods.go:89] "kube-controller-manager-addons-246651" [628c8d4b-3ef2-4ac9-9594-d919b039c856] Running
I0717 00:22:44.711196 12110 system_pods.go:89] "kube-ingress-dns-minikube" [0ed9119c-c7eb-4c2f-ba3c-79fea7e2ba91] Running
I0717 00:22:44.711202 12110 system_pods.go:89] "kube-proxy-nzcqj" [0d972af8-6f49-43d1-84ed-0e938d71043b] Running
I0717 00:22:44.711209 12110 system_pods.go:89] "kube-scheduler-addons-246651" [d6598e71-3c71-43b0-b05d-5f199d4e0c34] Running
I0717 00:22:44.711214 12110 system_pods.go:89] "metrics-server-c59844bb4-2vpkj" [e3e50235-5bcc-435e-b32a-3db12ef1208a] Pending / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0717 00:22:44.711222 12110 system_pods.go:89] "nvidia-device-plugin-daemonset-g6hlf" [8c31d1b6-e782-4497-b523-0002031aa0c8] Running
I0717 00:22:44.711226 12110 system_pods.go:89] "registry-mwnw8" [922f7c1c-219a-49d4-9764-6789d827e7e8] Running
I0717 00:22:44.711231 12110 system_pods.go:89] "registry-proxy-crj68" [23346467-00bb-4945-bb70-82048f3eb257] Pending / Ready:ContainersNotReady (containers with unready status: [registry-proxy]) / ContainersReady:ContainersNotReady (containers with unready status: [registry-proxy])
I0717 00:22:44.711239 12110 system_pods.go:89] "snapshot-controller-745499f584-x95mr" [71c7a587-246c-4445-b115-08a1d678211d] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0717 00:22:44.711245 12110 system_pods.go:89] "snapshot-controller-745499f584-xmg7r" [164aa91a-825f-4cbc-a1cf-96f7d7fc5391] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0717 00:22:44.711251 12110 system_pods.go:89] "storage-provisioner" [1518d476-d81f-4a67-aed7-5f0ead3ab81c] Running
I0717 00:22:44.711255 12110 system_pods.go:89] "tiller-deploy-6677d64bcd-6llsw" [d8908f47-fb82-48b4-881d-6bd342a17f86] Running
I0717 00:22:44.711262 12110 system_pods.go:126] duration metric: took 206.526044ms to wait for k8s-apps to be running ...
I0717 00:22:44.711272 12110 system_svc.go:44] waiting for kubelet service to be running ....
I0717 00:22:44.711323 12110 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0717 00:22:44.724825 12110 system_svc.go:56] duration metric: took 13.540175ms WaitForService to wait for kubelet
I0717 00:22:44.724860 12110 kubeadm.go:582] duration metric: took 41.238177143s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0717 00:22:44.724886 12110 node_conditions.go:102] verifying NodePressure condition ...
I0717 00:22:44.811226 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0717 00:22:44.895828 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:44.904731 12110 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0717 00:22:44.904759 12110 node_conditions.go:123] node cpu capacity is 8
I0717 00:22:44.904772 12110 node_conditions.go:105] duration metric: took 179.880542ms to run NodePressure ...
I0717 00:22:44.904787 12110 start.go:241] waiting for startup goroutines ...
I0717 00:22:45.091803 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:45.310361 12110 kapi.go:107] duration metric: took 30.004554747s to wait for kubernetes.io/minikube-addons=registry ...
I0717 00:22:45.395172 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:45.592424 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:45.896031 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:46.092734 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:46.395863 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:46.591871 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:46.897070 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:47.092144 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:47.394992 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:47.592955 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:47.895834 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:48.092047 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:48.396106 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:48.592980 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:48.896430 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:49.092867 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:49.396565 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:49.592900 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:49.896610 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:50.092530 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:50.396155 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:50.592666 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:51.048916 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:51.091810 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:51.396092 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:51.592815 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:51.895774 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:52.092384 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:52.396610 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:52.592432 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:52.896078 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:53.091745 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:53.396284 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:53.592182 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:53.896393 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:54.093004 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:54.396415 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:54.592872 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:54.896278 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:55.091802 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:55.397043 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:55.592277 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:55.896890 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:56.092071 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:56.396014 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:56.591366 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:56.895515 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:57.092351 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:57.395926 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:57.592704 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:57.896551 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:58.095234 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:58.395745 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:58.591783 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:58.895981 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:59.091548 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:59.395777 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:22:59.590982 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:22:59.895792 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:00.091836 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:00.395906 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:00.591921 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:00.896580 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:01.092825 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:01.395770 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:01.593988 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:01.895659 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:02.091943 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:02.395639 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:02.593211 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:02.897008 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:03.091769 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:03.395615 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:03.592008 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:03.896322 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:04.092870 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:04.394958 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:04.591339 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:04.896087 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:05.092640 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:05.395435 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:05.591791 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:05.896715 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:06.092937 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:06.416288 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:06.592149 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:06.895860 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:07.091929 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:07.397364 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:07.592378 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:07.896203 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:08.092804 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:08.396056 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:08.592695 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:08.896796 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:09.092310 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:09.396569 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:09.592385 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:09.895304 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:10.093359 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:10.396968 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:10.592722 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:10.896117 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:11.091512 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:11.395664 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:11.592209 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:11.896553 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:12.091819 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:12.395559 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:12.591185 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0717 00:23:12.895314 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:13.091583 12110 kapi.go:107] duration metric: took 56.504615851s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0717 00:23:13.395659 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:13.895888 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:14.396319 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:14.896081 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:15.395262 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:15.895685 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:16.395504 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:16.896279 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:17.395878 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:17.896281 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:18.396384 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:18.896977 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:19.395920 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:19.896641 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:20.395955 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:20.896793 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:21.396153 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:21.896189 12110 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0717 00:23:22.395059 12110 kapi.go:107] duration metric: took 1m10.003325607s to wait for app.kubernetes.io/name=ingress-nginx ...
I0717 00:23:41.218843 12110 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0717 00:23:41.218872 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:41.716091 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:42.216056 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:42.715782 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:43.217041 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:43.715827 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:44.216721 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:44.716460 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:45.216651 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:45.716183 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:46.216163 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:46.716047 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:47.215840 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:47.716636 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:48.216459 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:48.716443 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:49.216185 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:49.715983 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:50.216299 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:50.716372 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:51.216131 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:51.715932 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:52.216979 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:52.716806 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:53.216685 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:53.716548 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:54.216369 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:54.716245 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:55.216110 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:55.715766 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:56.216309 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:56.716088 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:57.215977 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:57.717129 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:58.216694 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:58.716508 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:59.216374 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:23:59.715805 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:00.216042 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:00.715676 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:01.216677 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:01.716418 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:02.216195 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:02.715885 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:03.216608 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:03.716514 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:04.216325 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:04.716782 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:05.216670 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:05.716190 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:06.215938 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:06.716709 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:07.216675 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:07.716545 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:08.216242 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:08.716066 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:09.215927 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:09.717265 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:10.216284 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:10.715939 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:11.216274 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:11.716027 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:12.216716 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:12.716340 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:13.216203 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:13.716105 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:14.216645 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:14.716229 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:15.216000 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:15.715642 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:16.216394 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:16.716091 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:17.216014 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:17.716708 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:18.216420 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:18.716149 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:19.215880 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:19.716592 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:20.216506 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:20.716281 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:21.218537 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:21.716657 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:22.216608 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:22.716362 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:23.216168 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:23.716108 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:24.216037 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:24.716915 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:25.216940 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:25.716931 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:26.216761 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:26.716623 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:27.216746 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:27.716943 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:28.216609 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:28.716483 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:29.216414 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:29.716219 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:30.216177 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:30.715960 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:31.216765 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:31.716613 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:32.216330 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:32.715960 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:33.216850 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:33.716614 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:34.216490 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:34.716102 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:35.216110 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:35.716115 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:36.215915 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:36.716345 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:37.215882 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:37.716774 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:38.216548 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:38.716152 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:39.216700 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:39.716773 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:40.216929 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:40.717187 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:41.216441 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:41.716847 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:42.218895 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:42.716652 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:43.216458 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:43.716603 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:44.216465 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:44.716250 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:45.216042 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:45.717939 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:46.215938 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:46.716566 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:47.216687 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:47.717225 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:48.215924 12110 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0717 00:24:48.717012 12110 kapi.go:107] duration metric: took 2m30.503847049s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0717 00:24:48.718668 12110 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-246651 cluster.
I0717 00:24:48.719844 12110 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0717 00:24:48.720954 12110 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0717 00:24:48.722011 12110 out.go:177] * Enabled addons: cloud-spanner, storage-provisioner, nvidia-device-plugin, storage-provisioner-rancher, ingress-dns, helm-tiller, default-storageclass, volcano, metrics-server, inspektor-gadget, yakd, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0717 00:24:48.723127 12110 addons.go:510] duration metric: took 2m45.236394541s for enable addons: enabled=[cloud-spanner storage-provisioner nvidia-device-plugin storage-provisioner-rancher ingress-dns helm-tiller default-storageclass volcano metrics-server inspektor-gadget yakd volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0717 00:24:48.723168 12110 start.go:246] waiting for cluster config update ...
I0717 00:24:48.723185 12110 start.go:255] writing updated cluster config ...
I0717 00:24:48.723421 12110 ssh_runner.go:195] Run: rm -f paused
I0717 00:24:48.769943 12110 start.go:600] kubectl: 1.30.2, cluster: 1.30.2 (minor skew: 0)
I0717 00:24:48.771760 12110 out.go:177] * Done! kubectl is now configured to use "addons-246651" cluster and "default" namespace by default
==> Docker <==
Jul 17 00:25:26 addons-246651 dockerd[1339]: time="2024-07-17T00:25:26.197267853Z" level=info msg="ignoring event" container=8996e36414829409db1d403efd10bcb5e9d2508d6b9dd57fa2f7b86f4f05b641 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:26 addons-246651 dockerd[1339]: time="2024-07-17T00:25:26.402876434Z" level=info msg="ignoring event" container=f193859bb77dea37e0103277d3a5befe65a446d8e44c514997eae5b669290bfd module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:26 addons-246651 dockerd[1339]: time="2024-07-17T00:25:26.686888181Z" level=info msg="ignoring event" container=7094c2df38f91069c7bd9014628f87faa229989f72bbf4067208eb4d7800d31c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:26 addons-246651 cri-dockerd[1603]: time="2024-07-17T00:25:26Z" level=info msg="Failed to read pod IP from plugin/docker: networkPlugin cni failed on the status hook for pod \"volcano-admission-5f7844f7bc-b452w_volcano-system\": unexpected command output nsenter: cannot open /proc/8914/ns/net: No such file or directory\n with error: exit status 1"
Jul 17 00:25:26 addons-246651 dockerd[1339]: time="2024-07-17T00:25:26.982786582Z" level=info msg="ignoring event" container=ec349b9741d6ffead13ee825ee34d5e52d46331a4bd70e01a42d9466810a2efa module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:27 addons-246651 dockerd[1339]: time="2024-07-17T00:25:27.005726921Z" level=info msg="ignoring event" container=5800c182486a70947ac05814d0db43244b4e3a17a084dada61c605455061ca2f module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:27 addons-246651 dockerd[1339]: time="2024-07-17T00:25:27.701595720Z" level=info msg="ignoring event" container=c5797e1fc89f3d4b93914c3053f199e9e191be9d88f474f4aca6877e0c62a56e module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:27 addons-246651 dockerd[1339]: time="2024-07-17T00:25:27.705980191Z" level=info msg="ignoring event" container=b6b1182358005a977483cc9d8f55b81c65d3131d23e2041362edfd4c88e284f7 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:28 addons-246651 cri-dockerd[1603]: time="2024-07-17T00:25:28Z" level=info msg="Stop pulling image docker.io/kicbase/echo-server:1.0: Status: Downloaded newer image for kicbase/echo-server:1.0"
Jul 17 00:25:30 addons-246651 cri-dockerd[1603]: time="2024-07-17T00:25:30Z" level=error msg="error getting RW layer size for container ID '3d8442b4602e198e48e7549d3d0c85df700269cf8e7b2ac0967a57a17004b972': Error response from daemon: No such container: 3d8442b4602e198e48e7549d3d0c85df700269cf8e7b2ac0967a57a17004b972"
Jul 17 00:25:30 addons-246651 cri-dockerd[1603]: time="2024-07-17T00:25:30Z" level=error msg="Set backoffDuration to : 1m0s for container ID '3d8442b4602e198e48e7549d3d0c85df700269cf8e7b2ac0967a57a17004b972'"
Jul 17 00:25:30 addons-246651 cri-dockerd[1603]: time="2024-07-17T00:25:30Z" level=error msg="error getting RW layer size for container ID '8996e36414829409db1d403efd10bcb5e9d2508d6b9dd57fa2f7b86f4f05b641': Error response from daemon: No such container: 8996e36414829409db1d403efd10bcb5e9d2508d6b9dd57fa2f7b86f4f05b641"
Jul 17 00:25:30 addons-246651 cri-dockerd[1603]: time="2024-07-17T00:25:30Z" level=error msg="Set backoffDuration to : 1m0s for container ID '8996e36414829409db1d403efd10bcb5e9d2508d6b9dd57fa2f7b86f4f05b641'"
Jul 17 00:25:31 addons-246651 dockerd[1339]: time="2024-07-17T00:25:31.806950004Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=a31fb62b2ac4dc763dde22c73e3eea1b9458ff7587c321f2d314de04882e2628
Jul 17 00:25:31 addons-246651 dockerd[1339]: time="2024-07-17T00:25:31.862992581Z" level=info msg="ignoring event" container=a31fb62b2ac4dc763dde22c73e3eea1b9458ff7587c321f2d314de04882e2628 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:31 addons-246651 dockerd[1339]: time="2024-07-17T00:25:31.986925567Z" level=info msg="ignoring event" container=73d03a1c565e23bdb88eaa1e5f70aea40164964d5fd383b045dd768707c3f6f2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:41 addons-246651 dockerd[1339]: time="2024-07-17T00:25:41.062890919Z" level=info msg="ignoring event" container=8ba5fcccf2aec7d997790f42d7c9b5a11f202f242accea1e6200ff93f5b96eef module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:46 addons-246651 dockerd[1339]: time="2024-07-17T00:25:46.749116734Z" level=info msg="Container failed to exit within 30s of signal 15 - using the force" container=e3ee1acaf48128cf443472b40228845c614c890736000b747f2edc92eb9c3d82
Jul 17 00:25:46 addons-246651 dockerd[1339]: time="2024-07-17T00:25:46.768571557Z" level=info msg="ignoring event" container=e3ee1acaf48128cf443472b40228845c614c890736000b747f2edc92eb9c3d82 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:46 addons-246651 dockerd[1339]: time="2024-07-17T00:25:46.879708882Z" level=info msg="ignoring event" container=c26a2d7b3b151cda98bd0028f70a6e2863d6943f6de6d56e72ffa125492b4f30 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:51 addons-246651 cri-dockerd[1603]: time="2024-07-17T00:25:51Z" level=error msg="error getting RW layer size for container ID '51c923e8cebe872502f89fcd624e1d4e16b128c04ddcd146db1397a1eaa6d905': Error response from daemon: No such container: 51c923e8cebe872502f89fcd624e1d4e16b128c04ddcd146db1397a1eaa6d905"
Jul 17 00:25:51 addons-246651 cri-dockerd[1603]: time="2024-07-17T00:25:51Z" level=error msg="Set backoffDuration to : 1m0s for container ID '51c923e8cebe872502f89fcd624e1d4e16b128c04ddcd146db1397a1eaa6d905'"
Jul 17 00:25:57 addons-246651 dockerd[1339]: time="2024-07-17T00:25:57.100694420Z" level=info msg="Container failed to exit within 30s of signal 3 - using the force" container=7987e42f4a7668486ce2ca35082d112ca3986e7e7279bf04f23481769b130021
Jul 17 00:25:57 addons-246651 dockerd[1339]: time="2024-07-17T00:25:57.119566882Z" level=info msg="ignoring event" container=7987e42f4a7668486ce2ca35082d112ca3986e7e7279bf04f23481769b130021 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Jul 17 00:25:57 addons-246651 dockerd[1339]: time="2024-07-17T00:25:57.247219720Z" level=info msg="ignoring event" container=57842cecc3c9df17341bf8ef1f121ee3131e8ec88113befe172163038817bd4c module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
39dd3b225bc9d kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 5 minutes ago Running hello-world-app 0 b99dd86bd6554 hello-world-app-6778b5fc9f-76nrg
4d496c77bdbbe nginx@sha256:a45ee5d042aaa9e81e013f97ae40c3dda26fbe98f22b6251acdf28e579560d55 6 minutes ago Running nginx 0 75b1a0d987b5e nginx
8f8eef7a9fdb7 ghcr.io/headlamp-k8s/headlamp@sha256:1c3f42aacd8eee1d3f1c63efb5a3b42da387ca1d87b77b0f486e8443201fcb37 6 minutes ago Running headlamp 0 639424d200080 headlamp-7867546754-w94j9
a3b5644971e4f gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 6 minutes ago Running gcp-auth 0 bab77475724db gcp-auth-5db96cd9b4-wztfb
f435d9dd08d4c registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f 8 minutes ago Running csi-snapshotter 0 89d8895a44bab csi-hostpathplugin-pn6m2
57b4a748834c9 registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 8 minutes ago Running csi-provisioner 0 89d8895a44bab csi-hostpathplugin-pn6m2
9f3ec01e2dfd0 registry.k8s.io/sig-storage/livenessprobe@sha256:cacee2b5c36dd59d4c7e8469c05c9e4ef53ecb2df9025fa8c10cdaf61bce62f0 8 minutes ago Running liveness-probe 0 89d8895a44bab csi-hostpathplugin-pn6m2
f3ec70048913a registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 8 minutes ago Running hostpath 0 89d8895a44bab csi-hostpathplugin-pn6m2
53ff1de55cf16 registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:f1c25991bac2fbb7f5fcf91ed9438df31e30edee6bed5a780464238aa09ad24c 8 minutes ago Running node-driver-registrar 0 89d8895a44bab csi-hostpathplugin-pn6m2
50a4066d5b51a registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 8 minutes ago Running csi-resizer 0 fbd05213ebd27 csi-hostpath-resizer-0
7b31c9145ab02 registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b 8 minutes ago Running csi-attacher 0 ad8a76fc80741 csi-hostpath-attacher-0
d6392c9599d6e registry.k8s.io/sig-storage/csi-external-health-monitor-controller@sha256:80b9ba94aa2afe24553d69bd165a6a51552d1582d68618ec00d3b804a7d9193c 8 minutes ago Running csi-external-health-monitor-controller 0 89d8895a44bab csi-hostpathplugin-pn6m2
8f5f8665bfc9a registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 8 minutes ago Running volume-snapshot-controller 0 201e6598d52a1 snapshot-controller-745499f584-xmg7r
d54eae2b4a574 registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 8 minutes ago Running volume-snapshot-controller 0 c647ac3bbc661 snapshot-controller-745499f584-x95mr
4b5fe387688a1 marcnuri/yakd@sha256:c5414196116a2266ad097b0468833b73ef1d6c7922241115fe203fb826381624 8 minutes ago Running yakd 0 31acb16e523c0 yakd-dashboard-799879c74f-qrmkq
9788e98170d78 6e38f40d628db 9 minutes ago Running storage-provisioner 0 184eaed24faac storage-provisioner
487fb07579280 cbb01a7bd410d 9 minutes ago Running coredns 0 3f0aa00647008 coredns-7db6d8ff4d-2hm4d
e31ee4a147778 53c535741fb44 9 minutes ago Running kube-proxy 0 02a2e89648808 kube-proxy-nzcqj
1a942b201d9cb 7820c83aa1394 9 minutes ago Running kube-scheduler 0 ca038f4409e29 kube-scheduler-addons-246651
fbb0e27b1ae9f 3861cfcd7c04c 9 minutes ago Running etcd 0 fdfac4f22d9f6 etcd-addons-246651
af50148e095e2 e874818b3caac 9 minutes ago Running kube-controller-manager 0 233c0506e32c1 kube-controller-manager-addons-246651
c5c5ed4be6fb7 56ce0fd9fb532 9 minutes ago Running kube-apiserver 0 a4455bc588aeb kube-apiserver-addons-246651
==> coredns [487fb0757928] <==
[INFO] 10.244.0.22:44108 - 62002 "A IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.071693728s
[INFO] 10.244.0.22:49899 - 31904 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.072384101s
[INFO] 10.244.0.22:36642 - 25929 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000081845s
[INFO] 10.244.0.22:49899 - 20692 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000078015s
[INFO] 10.244.0.22:35914 - 7397 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000052267s
[INFO] 10.244.0.22:49272 - 30516 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00007478s
[INFO] 10.244.0.22:43194 - 29938 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.007327396s
[INFO] 10.244.0.22:48912 - 54344 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.006411086s
[INFO] 10.244.0.22:43194 - 39126 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000071197s
[INFO] 10.244.0.22:54799 - 9216 "AAAA IN hello-world-app.default.svc.cluster.local.europe-west2-a.c.k8s-minikube.internal. udp 98 false 512" NXDOMAIN qr,rd,ra 98 0.006753441s
[INFO] 10.244.0.22:48912 - 7399 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000112549s
[INFO] 10.244.0.22:44108 - 24091 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.002949689s
[INFO] 10.244.0.22:35524 - 19815 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003890747s
[INFO] 10.244.0.22:54799 - 56234 "A IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.002722542s
[INFO] 10.244.0.22:35524 - 41590 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000056602s
[INFO] 10.244.0.22:44108 - 14329 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.001778075s
[INFO] 10.244.0.22:49598 - 6042 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004989198s
[INFO] 10.244.0.22:54799 - 63066 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.003004987s
[INFO] 10.244.0.22:44108 - 31534 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005600161s
[INFO] 10.244.0.22:49598 - 6755 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005636002s
[INFO] 10.244.0.22:44108 - 14949 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000038964s
[INFO] 10.244.0.22:49598 - 51263 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00002563s
[INFO] 10.244.0.22:54799 - 559 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.003973115s
[INFO] 10.244.0.22:54799 - 54255 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.002909995s
[INFO] 10.244.0.22:54799 - 57394 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000058065s
==> describe nodes <==
Name: addons-246651
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-246651
kubernetes.io/os=linux
minikube.k8s.io/commit=3cfbbb17fd76400a5ee2ea427db7148a0ef7c185
minikube.k8s.io/name=addons-246651
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_07_17T00_21_49_0700
minikube.k8s.io/version=v1.33.1
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-246651
Annotations: csi.volume.kubernetes.io/nodeid: {"hostpath.csi.k8s.io":"addons-246651"}
kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Wed, 17 Jul 2024 00:21:46 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-246651
AcquireTime: <unset>
RenewTime: Wed, 17 Jul 2024 00:31:20 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Wed, 17 Jul 2024 00:31:00 +0000 Wed, 17 Jul 2024 00:21:44 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Wed, 17 Jul 2024 00:31:00 +0000 Wed, 17 Jul 2024 00:21:44 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Wed, 17 Jul 2024 00:31:00 +0000 Wed, 17 Jul 2024 00:21:44 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Wed, 17 Jul 2024 00:31:00 +0000 Wed, 17 Jul 2024 00:21:46 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-246651
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859328Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859328Ki
pods: 110
System Info:
Machine ID: edfbf2d63ffe4a6691a910c9af51dd8b
System UUID: e2643a7f-ce81-4513-81e9-ed218f9b924f
Boot ID: 2ca48bdf-eb97-45ac-92d1-65a3de076a8d
Kernel Version: 5.15.0-1062-gcp
OS Image: Ubuntu 22.04.4 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.0.3
Kubelet Version: v1.30.2
Kube-Proxy Version: v1.30.2
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (17 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default hello-world-app-6778b5fc9f-76nrg 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 6m3s
default nginx 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 6m14s
gcp-auth gcp-auth-5db96cd9b4-wztfb 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 7m47s
headlamp headlamp-7867546754-w94j9 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 6m39s
kube-system coredns-7db6d8ff4d-2hm4d 100m (1%!)(MISSING) 0 (0%!)(MISSING) 70Mi (0%!)(MISSING) 170Mi (0%!)(MISSING) 9m26s
kube-system csi-hostpath-attacher-0 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 9m12s
kube-system csi-hostpath-resizer-0 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 9m12s
kube-system csi-hostpathplugin-pn6m2 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 9m12s
kube-system etcd-addons-246651 100m (1%!)(MISSING) 0 (0%!)(MISSING) 100Mi (0%!)(MISSING) 0 (0%!)(MISSING) 9m40s
kube-system kube-apiserver-addons-246651 250m (3%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 9m40s
kube-system kube-controller-manager-addons-246651 200m (2%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 9m40s
kube-system kube-proxy-nzcqj 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 9m26s
kube-system kube-scheduler-addons-246651 100m (1%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 9m40s
kube-system snapshot-controller-745499f584-x95mr 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 9m16s
kube-system snapshot-controller-745499f584-xmg7r 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 9m16s
kube-system storage-provisioner 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 0 (0%!)(MISSING) 9m21s
yakd-dashboard yakd-dashboard-799879c74f-qrmkq 0 (0%!)(MISSING) 0 (0%!)(MISSING) 128Mi (0%!)(MISSING) 256Mi (0%!)(MISSING) 9m18s
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%!)(MISSING) 0 (0%!)(MISSING)
memory 298Mi (0%!)(MISSING) 426Mi (1%!)(MISSING)
ephemeral-storage 0 (0%!)(MISSING) 0 (0%!)(MISSING)
hugepages-1Gi 0 (0%!)(MISSING) 0 (0%!)(MISSING)
hugepages-2Mi 0 (0%!)(MISSING) 0 (0%!)(MISSING)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 9m24s kube-proxy
Normal Starting 9m40s kubelet Starting kubelet.
Normal NodeAllocatableEnforced 9m40s kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 9m40s kubelet Node addons-246651 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 9m40s kubelet Node addons-246651 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 9m40s kubelet Node addons-246651 status is now: NodeHasSufficientPID
Normal RegisteredNode 9m26s node-controller Node addons-246651 event: Registered Node addons-246651 in Controller
==> dmesg <==
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff c6 cd 0c d7 52 2f 08 06
[ +12.611765] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 76 e6 ea 24 09 ed 08 06
[ +1.080780] IPv4: martian source 10.244.0.1 from 10.244.0.23, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff d6 df ed ea 07 96 08 06
[Jul17 00:24] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff ae c2 e0 9f eb 81 08 06
[ +0.018984] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 42 e6 1f 6d dc 7f 08 06
[ +26.203290] IPv4: martian source 10.244.0.1 from 10.244.0.26, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff de 4f 6c f2 67 1e 08 06
[ +0.000434] IPv4: martian source 10.244.0.26 from 10.244.0.3, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff aa 95 01 cf 18 eb 08 06
[ +8.516591] IPv4: martian source 10.244.0.1 from 10.244.0.28, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff ce 55 22 ba 2f 1c 08 06
[Jul17 00:25] IPv4: martian source 10.244.0.1 from 10.244.0.29, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff 8e 04 58 c5 b1 56 08 06
[ +0.000400] IPv4: martian source 10.244.0.29 from 10.244.0.3, on dev eth0
[ +0.000003] ll header: 00000000: ff ff ff ff ff ff aa 95 01 cf 18 eb 08 06
[ +0.000523] IPv4: martian source 10.244.0.29 from 10.244.0.8, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff 92 62 fc 18 a3 d7 08 06
[ +23.822612] IPv4: martian source 10.244.0.33 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 76 e6 ea 24 09 ed 08 06
[ +3.030038] IPv4: martian source 10.244.0.22 from 10.244.0.3, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff aa 95 01 cf 18 eb 08 06
==> etcd [fbb0e27b1ae9] <==
{"level":"info","ts":"2024-07-17T00:21:44.709643Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-07-17T00:21:44.709657Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-07-17T00:21:44.70967Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-07-17T00:21:44.710561Z","caller":"etcdserver/server.go:2068","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-246651 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-07-17T00:21:44.710607Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-07-17T00:21:44.710665Z","caller":"etcdserver/server.go:2578","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-07-17T00:21:44.71073Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-07-17T00:21:44.710741Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-07-17T00:21:44.710701Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-07-17T00:21:44.711303Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-07-17T00:21:44.711393Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-07-17T00:21:44.711414Z","caller":"etcdserver/server.go:2602","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-07-17T00:21:44.712738Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-07-17T00:21:44.713166Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-07-17T00:22:21.90985Z","caller":"traceutil/trace.go:171","msg":"trace[1374413786] transaction","detail":"{read_only:false; response_revision:968; number_of_response:1; }","duration":"109.294178ms","start":"2024-07-17T00:22:21.800543Z","end":"2024-07-17T00:22:21.909837Z","steps":["trace[1374413786] 'process raft request' (duration: 109.180496ms)"],"step_count":1}
{"level":"info","ts":"2024-07-17T00:22:21.909839Z","caller":"traceutil/trace.go:171","msg":"trace[764716699] linearizableReadLoop","detail":"{readStateIndex:989; appliedIndex:988; }","duration":"102.079536ms","start":"2024-07-17T00:22:21.807742Z","end":"2024-07-17T00:22:21.909821Z","steps":["trace[764716699] 'read index received' (duration: 101.931643ms)","trace[764716699] 'applied index is now lower than readState.Index' (duration: 147.271µs)"],"step_count":2}
{"level":"warn","ts":"2024-07-17T00:22:21.910121Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"102.323129ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods/kube-system/\" range_end:\"/registry/pods/kube-system0\" ","response":"range_response_count:18 size:86954"}
{"level":"info","ts":"2024-07-17T00:22:21.910195Z","caller":"traceutil/trace.go:171","msg":"trace[1700128140] range","detail":"{range_begin:/registry/pods/kube-system/; range_end:/registry/pods/kube-system0; response_count:18; response_revision:968; }","duration":"102.465273ms","start":"2024-07-17T00:22:21.807717Z","end":"2024-07-17T00:22:21.910182Z","steps":["trace[1700128140] 'agreement among raft nodes before linearized reading' (duration: 102.172477ms)"],"step_count":1}
{"level":"info","ts":"2024-07-17T00:22:50.783696Z","caller":"traceutil/trace.go:171","msg":"trace[411690290] transaction","detail":"{read_only:false; response_revision:1073; number_of_response:1; }","duration":"133.745642ms","start":"2024-07-17T00:22:50.649924Z","end":"2024-07-17T00:22:50.78367Z","steps":["trace[411690290] 'process raft request' (duration: 66.52789ms)","trace[411690290] 'compare' (duration: 67.101405ms)"],"step_count":2}
{"level":"warn","ts":"2024-07-17T00:22:51.046904Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"152.761045ms","expected-duration":"100ms","prefix":"read-only range ","request":"key:\"/registry/pods/ingress-nginx/\" range_end:\"/registry/pods/ingress-nginx0\" ","response":"range_response_count:3 size:14065"}
{"level":"info","ts":"2024-07-17T00:22:51.046975Z","caller":"traceutil/trace.go:171","msg":"trace[1626180266] range","detail":"{range_begin:/registry/pods/ingress-nginx/; range_end:/registry/pods/ingress-nginx0; response_count:3; response_revision:1074; }","duration":"152.869406ms","start":"2024-07-17T00:22:50.894092Z","end":"2024-07-17T00:22:51.046961Z","steps":["trace[1626180266] 'range keys from in-memory index tree' (duration: 152.59885ms)"],"step_count":1}
{"level":"warn","ts":"2024-07-17T00:23:21.534652Z","caller":"etcdserver/util.go:170","msg":"apply request took too long","took":"112.106647ms","expected-duration":"100ms","prefix":"read-only range ","request":"limit:1 serializable:true keys_only:true ","response":"range_response_count:0 size:5"}
{"level":"info","ts":"2024-07-17T00:23:21.534752Z","caller":"traceutil/trace.go:171","msg":"trace[46221787] range","detail":"{range_begin:; range_end:; response_count:0; response_revision:1254; }","duration":"112.221534ms","start":"2024-07-17T00:23:21.422517Z","end":"2024-07-17T00:23:21.534739Z","steps":["trace[46221787] 'range keys from in-memory index tree' (duration: 112.092149ms)"],"step_count":1}
{"level":"info","ts":"2024-07-17T00:23:21.569633Z","caller":"traceutil/trace.go:171","msg":"trace[1041144901] transaction","detail":"{read_only:false; response_revision:1255; number_of_response:1; }","duration":"111.012804ms","start":"2024-07-17T00:23:21.458601Z","end":"2024-07-17T00:23:21.569614Z","steps":["trace[1041144901] 'process raft request' (duration: 110.912726ms)"],"step_count":1}
{"level":"info","ts":"2024-07-17T00:23:21.812687Z","caller":"traceutil/trace.go:171","msg":"trace[190344255] transaction","detail":"{read_only:false; response_revision:1256; number_of_response:1; }","duration":"198.857373ms","start":"2024-07-17T00:23:21.613805Z","end":"2024-07-17T00:23:21.812663Z","steps":["trace[190344255] 'process raft request' (duration: 125.345793ms)","trace[190344255] 'compare' (duration: 73.360785ms)"],"step_count":2}
==> gcp-auth [a3b5644971e4] <==
2024/07/17 00:24:49 Ready to write response ...
2024/07/17 00:24:49 Ready to marshal response ...
2024/07/17 00:24:49 Ready to write response ...
2024/07/17 00:24:53 Ready to marshal response ...
2024/07/17 00:24:53 Ready to write response ...
2024/07/17 00:24:58 Ready to marshal response ...
2024/07/17 00:24:58 Ready to write response ...
2024/07/17 00:25:01 Ready to marshal response ...
2024/07/17 00:25:01 Ready to write response ...
2024/07/17 00:25:01 Ready to marshal response ...
2024/07/17 00:25:01 Ready to write response ...
2024/07/17 00:25:05 Ready to marshal response ...
2024/07/17 00:25:05 Ready to write response ...
2024/07/17 00:25:14 Ready to marshal response ...
2024/07/17 00:25:14 Ready to write response ...
2024/07/17 00:25:15 Ready to marshal response ...
2024/07/17 00:25:15 Ready to write response ...
2024/07/17 00:25:15 Ready to marshal response ...
2024/07/17 00:25:15 Ready to write response ...
2024/07/17 00:25:16 Ready to marshal response ...
2024/07/17 00:25:16 Ready to write response ...
2024/07/17 00:25:25 Ready to marshal response ...
2024/07/17 00:25:25 Ready to write response ...
2024/07/17 00:25:26 Ready to marshal response ...
2024/07/17 00:25:26 Ready to write response ...
==> kernel <==
00:31:28 up 13 min, 0 users, load average: 0.05, 0.33, 0.32
Linux addons-246651 5.15.0-1062-gcp #70~20.04.1-Ubuntu SMP Fri May 24 20:12:18 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.4 LTS"
==> kube-apiserver [c5c5ed4be6fb] <==
I0717 00:25:14.928822 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.100.11.12"}
I0717 00:25:15.535455 1 controller.go:615] quota admission added evaluator for: jobs.batch.volcano.sh
I0717 00:25:15.551304 1 controller.go:615] quota admission added evaluator for: podgroups.scheduling.volcano.sh
I0717 00:25:17.799555 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
I0717 00:25:25.404238 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.108.192.28"}
I0717 00:25:26.300201 1 handler.go:286] Adding GroupVersion batch.volcano.sh v1alpha1 to ResourceManager
I0717 00:25:26.383514 1 handler.go:286] Adding GroupVersion bus.volcano.sh v1alpha1 to ResourceManager
I0717 00:25:26.697896 1 handler.go:286] Adding GroupVersion scheduling.volcano.sh v1beta1 to ResourceManager
I0717 00:25:26.698579 1 handler.go:286] Adding GroupVersion scheduling.volcano.sh v1beta1 to ResourceManager
I0717 00:25:26.705383 1 handler.go:286] Adding GroupVersion nodeinfo.volcano.sh v1alpha1 to ResourceManager
W0717 00:25:26.893214 1 dispatcher.go:217] Failed calling webhook, failing closed validatepod.volcano.sh: failed calling webhook "validatepod.volcano.sh": failed to call webhook: Post "https://volcano-admission-service.volcano-system.svc:443/pods/validate?timeout=10s": service "volcano-admission-service" not found
I0717 00:25:27.082961 1 handler.go:286] Adding GroupVersion scheduling.volcano.sh v1beta1 to ResourceManager
I0717 00:25:27.383206 1 handler.go:286] Adding GroupVersion flow.volcano.sh v1alpha1 to ResourceManager
W0717 00:25:27.408792 1 cacher.go:168] Terminating all watchers from cacher commands.bus.volcano.sh
I0717 00:25:27.485969 1 handler.go:286] Adding GroupVersion flow.volcano.sh v1alpha1 to ResourceManager
I0717 00:25:27.603338 1 handler.go:286] Adding GroupVersion flow.volcano.sh v1alpha1 to ResourceManager
W0717 00:25:28.003216 1 cacher.go:168] Terminating all watchers from cacher numatopologies.nodeinfo.volcano.sh
W0717 00:25:28.083219 1 cacher.go:168] Terminating all watchers from cacher podgroups.scheduling.volcano.sh
W0717 00:25:28.098814 1 cacher.go:168] Terminating all watchers from cacher queues.scheduling.volcano.sh
W0717 00:25:28.301077 1 cacher.go:168] Terminating all watchers from cacher jobs.batch.volcano.sh
W0717 00:25:28.603538 1 cacher.go:168] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0717 00:25:29.299948 1 cacher.go:168] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
E0717 00:25:32.256841 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
I0717 00:25:41.018187 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0717 00:25:42.031544 1 cacher.go:168] Terminating all watchers from cacher traces.gadget.kinvolk.io
==> kube-controller-manager [af50148e095e] <==
E0717 00:30:09.257789 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:30:11.838791 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:30:11.838829 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:30:18.405319 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:30:18.405362 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:30:31.787960 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:30:31.787995 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:30:32.422055 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:30:32.422116 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:30:36.468461 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:30:36.468493 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:30:51.845619 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:30:51.845658 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:30:52.429584 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:30:52.429623 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:30:52.867485 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:30:52.867517 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:30:53.898888 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:30:53.898921 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:31:03.190827 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:31:03.190869 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:31:10.098052 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:31:10.098114 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
W0717 00:31:20.757039 1 reflector.go:547] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0717 00:31:20.757069 1 reflector.go:150] k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
==> kube-proxy [e31ee4a14777] <==
I0717 00:22:03.392742 1 server_linux.go:69] "Using iptables proxy"
I0717 00:22:03.401958 1 server.go:1062] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
I0717 00:22:03.483731 1 server.go:659] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0717 00:22:03.483792 1 server_linux.go:165] "Using iptables Proxier"
I0717 00:22:03.488523 1 server_linux.go:511] "Detect-local-mode set to ClusterCIDR, but no cluster CIDR for family" ipFamily="IPv6"
I0717 00:22:03.488548 1 server_linux.go:528] "Defaulting to no-op detect-local"
I0717 00:22:03.488579 1 proxier.go:243] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses"
I0717 00:22:03.488830 1 server.go:872] "Version info" version="v1.30.2"
I0717 00:22:03.489015 1 server.go:874] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0717 00:22:03.492408 1 config.go:192] "Starting service config controller"
I0717 00:22:03.493642 1 shared_informer.go:313] Waiting for caches to sync for service config
I0717 00:22:03.492604 1 config.go:319] "Starting node config controller"
I0717 00:22:03.493887 1 shared_informer.go:313] Waiting for caches to sync for node config
I0717 00:22:03.493422 1 config.go:101] "Starting endpoint slice config controller"
I0717 00:22:03.493970 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0717 00:22:03.493997 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0717 00:22:03.595798 1 shared_informer.go:320] Caches are synced for node config
I0717 00:22:03.595829 1 shared_informer.go:320] Caches are synced for service config
==> kube-scheduler [1a942b201d9c] <==
W0717 00:21:46.595915 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0717 00:21:46.595930 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
W0717 00:21:47.402305 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
E0717 00:21:47.402335 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope
W0717 00:21:47.478522 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0717 00:21:47.478553 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
W0717 00:21:47.488511 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
E0717 00:21:47.488539 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
W0717 00:21:47.491372 1 reflector.go:547] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0717 00:21:47.491394 1 reflector.go:150] runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
W0717 00:21:47.549776 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0717 00:21:47.549814 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
W0717 00:21:47.602284 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
E0717 00:21:47.602317 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
W0717 00:21:47.651239 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
E0717 00:21:47.651276 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope
W0717 00:21:47.663203 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
E0717 00:21:47.663240 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
W0717 00:21:47.680394 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
E0717 00:21:47.680435 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StatefulSet: failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope
W0717 00:21:47.680401 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0717 00:21:47.680479 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
W0717 00:21:47.711889 1 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0717 00:21:47.711922 1 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
I0717 00:21:50.593206 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Jul 17 00:25:47 addons-246651 kubelet[2537]: I0717 00:25:47.122011 2537 reconciler_common.go:289] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e714ff18-3051-4d1c-b758-4bb8e5533505-config-volume\") on node \"addons-246651\" DevicePath \"\""
Jul 17 00:25:47 addons-246651 kubelet[2537]: I0717 00:25:47.612896 2537 scope.go:117] "RemoveContainer" containerID="e3ee1acaf48128cf443472b40228845c614c890736000b747f2edc92eb9c3d82"
Jul 17 00:25:47 addons-246651 kubelet[2537]: I0717 00:25:47.625940 2537 scope.go:117] "RemoveContainer" containerID="e3ee1acaf48128cf443472b40228845c614c890736000b747f2edc92eb9c3d82"
Jul 17 00:25:47 addons-246651 kubelet[2537]: E0717 00:25:47.626826 2537 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: e3ee1acaf48128cf443472b40228845c614c890736000b747f2edc92eb9c3d82" containerID="e3ee1acaf48128cf443472b40228845c614c890736000b747f2edc92eb9c3d82"
Jul 17 00:25:47 addons-246651 kubelet[2537]: I0717 00:25:47.626866 2537 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"e3ee1acaf48128cf443472b40228845c614c890736000b747f2edc92eb9c3d82"} err="failed to get container status \"e3ee1acaf48128cf443472b40228845c614c890736000b747f2edc92eb9c3d82\": rpc error: code = Unknown desc = Error response from daemon: No such container: e3ee1acaf48128cf443472b40228845c614c890736000b747f2edc92eb9c3d82"
Jul 17 00:25:48 addons-246651 kubelet[2537]: I0717 00:25:48.898139 2537 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e714ff18-3051-4d1c-b758-4bb8e5533505" path="/var/lib/kubelet/pods/e714ff18-3051-4d1c-b758-4bb8e5533505/volumes"
Jul 17 00:25:48 addons-246651 kubelet[2537]: I0717 00:25:48.964828 2537 scope.go:117] "RemoveContainer" containerID="c87e5d60c874e44540484cb3e0ca11cacdac25a6914429363b8f08dea9487745"
Jul 17 00:25:48 addons-246651 kubelet[2537]: I0717 00:25:48.977187 2537 scope.go:117] "RemoveContainer" containerID="5e7a0bf51c3ca8e85ca1ea5ae97febfce0589e637d052950f1d4d60bc6eb1393"
Jul 17 00:25:48 addons-246651 kubelet[2537]: I0717 00:25:48.988516 2537 scope.go:117] "RemoveContainer" containerID="fdb004d65fa2a2a4a4b90d752fe997b982644f6bcbdf3692b64eb369a29af5fd"
Jul 17 00:25:49 addons-246651 kubelet[2537]: I0717 00:25:49.003304 2537 scope.go:117] "RemoveContainer" containerID="ef81672083e2d740158898fdc7956153221bf896313f8cf19f775f7cd7cd727c"
Jul 17 00:25:49 addons-246651 kubelet[2537]: I0717 00:25:49.014663 2537 scope.go:117] "RemoveContainer" containerID="e7fe2db272755b24e5164f41704d48581f9b6929fd37b45fa60b0d7283fecef0"
Jul 17 00:25:49 addons-246651 kubelet[2537]: I0717 00:25:49.025670 2537 scope.go:117] "RemoveContainer" containerID="f816db35be92662eef16aab770e65fa329e122cddc009dbf69a0fa3b33b9acec"
Jul 17 00:25:49 addons-246651 kubelet[2537]: I0717 00:25:49.039077 2537 scope.go:117] "RemoveContainer" containerID="be7ef78f6c22960fa8483d5ca1851c5ffabe541ff766cc3618e43be237bf4780"
Jul 17 00:25:49 addons-246651 kubelet[2537]: I0717 00:25:49.051924 2537 scope.go:117] "RemoveContainer" containerID="5e67c3d823c3af4a6ff854b4a0fbfd2f7726cc524dddbf6f6a6b05699d512e28"
Jul 17 00:25:57 addons-246651 kubelet[2537]: I0717 00:25:57.383148 2537 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/57149743-b99d-4641-a579-cd5c54a6dea6-gcp-creds\") pod \"57149743-b99d-4641-a579-cd5c54a6dea6\" (UID: \"57149743-b99d-4641-a579-cd5c54a6dea6\") "
Jul 17 00:25:57 addons-246651 kubelet[2537]: I0717 00:25:57.383211 2537 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7n5b\" (UniqueName: \"kubernetes.io/projected/57149743-b99d-4641-a579-cd5c54a6dea6-kube-api-access-q7n5b\") pod \"57149743-b99d-4641-a579-cd5c54a6dea6\" (UID: \"57149743-b99d-4641-a579-cd5c54a6dea6\") "
Jul 17 00:25:57 addons-246651 kubelet[2537]: I0717 00:25:57.383270 2537 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57149743-b99d-4641-a579-cd5c54a6dea6-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "57149743-b99d-4641-a579-cd5c54a6dea6" (UID: "57149743-b99d-4641-a579-cd5c54a6dea6"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul 17 00:25:57 addons-246651 kubelet[2537]: I0717 00:25:57.384914 2537 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57149743-b99d-4641-a579-cd5c54a6dea6-kube-api-access-q7n5b" (OuterVolumeSpecName: "kube-api-access-q7n5b") pod "57149743-b99d-4641-a579-cd5c54a6dea6" (UID: "57149743-b99d-4641-a579-cd5c54a6dea6"). InnerVolumeSpecName "kube-api-access-q7n5b". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jul 17 00:25:57 addons-246651 kubelet[2537]: I0717 00:25:57.483976 2537 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-q7n5b\" (UniqueName: \"kubernetes.io/projected/57149743-b99d-4641-a579-cd5c54a6dea6-kube-api-access-q7n5b\") on node \"addons-246651\" DevicePath \"\""
Jul 17 00:25:57 addons-246651 kubelet[2537]: I0717 00:25:57.484013 2537 reconciler_common.go:289] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/57149743-b99d-4641-a579-cd5c54a6dea6-gcp-creds\") on node \"addons-246651\" DevicePath \"\""
Jul 17 00:25:57 addons-246651 kubelet[2537]: I0717 00:25:57.720054 2537 scope.go:117] "RemoveContainer" containerID="7987e42f4a7668486ce2ca35082d112ca3986e7e7279bf04f23481769b130021"
Jul 17 00:25:57 addons-246651 kubelet[2537]: I0717 00:25:57.733356 2537 scope.go:117] "RemoveContainer" containerID="7987e42f4a7668486ce2ca35082d112ca3986e7e7279bf04f23481769b130021"
Jul 17 00:25:57 addons-246651 kubelet[2537]: E0717 00:25:57.734052 2537 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 7987e42f4a7668486ce2ca35082d112ca3986e7e7279bf04f23481769b130021" containerID="7987e42f4a7668486ce2ca35082d112ca3986e7e7279bf04f23481769b130021"
Jul 17 00:25:57 addons-246651 kubelet[2537]: I0717 00:25:57.734121 2537 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"7987e42f4a7668486ce2ca35082d112ca3986e7e7279bf04f23481769b130021"} err="failed to get container status \"7987e42f4a7668486ce2ca35082d112ca3986e7e7279bf04f23481769b130021\": rpc error: code = Unknown desc = Error response from daemon: No such container: 7987e42f4a7668486ce2ca35082d112ca3986e7e7279bf04f23481769b130021"
Jul 17 00:25:58 addons-246651 kubelet[2537]: I0717 00:25:58.897049 2537 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57149743-b99d-4641-a579-cd5c54a6dea6" path="/var/lib/kubelet/pods/57149743-b99d-4641-a579-cd5c54a6dea6/volumes"
==> storage-provisioner [9788e98170d7] <==
I0717 00:22:10.096510 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0717 00:22:10.188042 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0717 00:22:10.188089 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0717 00:22:10.196979 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0717 00:22:10.197162 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-246651_445ab121-8ef5-4b34-a4ad-8f5dc7335b46!
I0717 00:22:10.198362 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"dde75700-0cde-45f4-a36e-48b3958fc058", APIVersion:"v1", ResourceVersion:"541", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-246651_445ab121-8ef5-4b34-a4ad-8f5dc7335b46 became leader
I0717 00:22:10.401564 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-246651_445ab121-8ef5-4b34-a4ad-8f5dc7335b46!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-246651 -n addons-246651
helpers_test.go:261: (dbg) Run: kubectl --context addons-246651 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:285: <<< TestAddons/parallel/CSI FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/CSI (389.38s)