=== RUN TestAddons/parallel/Registry
=== PAUSE TestAddons/parallel/Registry
=== CONT TestAddons/parallel/Registry
addons_test.go:332: registry stabilized in 1.838181ms
addons_test.go:334: (dbg) TestAddons/parallel/Registry: waiting 6m0s for pods matching "actual-registry=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-6fb4cdfc84-d989j" [03e8d3f6-b271-4c1d-a933-77656192d8eb] Running
addons_test.go:334: (dbg) TestAddons/parallel/Registry: actual-registry=true healthy within 5.003492499s
addons_test.go:337: (dbg) TestAddons/parallel/Registry: waiting 10m0s for pods matching "registry-proxy=true" in namespace "kube-system" ...
helpers_test.go:344: "registry-proxy-c7665" [23346424-7dc9-4199-af57-8fffc4bcfe51] Running
addons_test.go:337: (dbg) TestAddons/parallel/Registry: registry-proxy=true healthy within 5.003384426s
addons_test.go:342: (dbg) Run: kubectl --context addons-141392 delete po -l run=registry-test --now
addons_test.go:347: (dbg) Run: kubectl --context addons-141392 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local"
addons_test.go:347: (dbg) Non-zero exit: kubectl --context addons-141392 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c "wget --spider -S http://registry.kube-system.svc.cluster.local": exit status 1 (1m0.07375626s)
-- stdout --
pod "registry-test" deleted
-- /stdout --
** stderr **
error: timed out waiting for the condition
** /stderr **
addons_test.go:349: failed to hit registry.kube-system.svc.cluster.local. args "kubectl --context addons-141392 run --rm registry-test --restart=Never --image=gcr.io/k8s-minikube/busybox -it -- sh -c \"wget --spider -S http://registry.kube-system.svc.cluster.local\"" failed: exit status 1
addons_test.go:353: expected curl response be "HTTP/1.1 200", but got *pod "registry-test" deleted
*
addons_test.go:361: (dbg) Run: out/minikube-linux-amd64 -p addons-141392 ip
2024/09/06 18:42:36 [DEBUG] GET http://192.168.49.2:5000
addons_test.go:390: (dbg) Run: out/minikube-linux-amd64 -p addons-141392 addons disable registry --alsologtostderr -v=1
helpers_test.go:222: -----------------------post-mortem--------------------------------
helpers_test.go:230: ======> post-mortem[TestAddons/parallel/Registry]: docker inspect <======
helpers_test.go:231: (dbg) Run: docker inspect addons-141392
helpers_test.go:235: (dbg) docker inspect addons-141392:
-- stdout --
[
{
"Id": "f37408764470cebc581a50c1543edaa2ff86fc719d5d0cb5050c14cbfbfa5b17",
"Created": "2024-09-06T18:29:42.599446993Z",
"Path": "/usr/local/bin/entrypoint",
"Args": [
"/sbin/init"
],
"State": {
"Status": "running",
"Running": true,
"Paused": false,
"Restarting": false,
"OOMKilled": false,
"Dead": false,
"Pid": 14666,
"ExitCode": 0,
"Error": "",
"StartedAt": "2024-09-06T18:29:42.71868642Z",
"FinishedAt": "0001-01-01T00:00:00Z"
},
"Image": "sha256:aeed0e1d4642008f872cbedd0f6935323c1e533683c40e800e0b01d063d11a3c",
"ResolvConfPath": "/var/lib/docker/containers/f37408764470cebc581a50c1543edaa2ff86fc719d5d0cb5050c14cbfbfa5b17/resolv.conf",
"HostnamePath": "/var/lib/docker/containers/f37408764470cebc581a50c1543edaa2ff86fc719d5d0cb5050c14cbfbfa5b17/hostname",
"HostsPath": "/var/lib/docker/containers/f37408764470cebc581a50c1543edaa2ff86fc719d5d0cb5050c14cbfbfa5b17/hosts",
"LogPath": "/var/lib/docker/containers/f37408764470cebc581a50c1543edaa2ff86fc719d5d0cb5050c14cbfbfa5b17/f37408764470cebc581a50c1543edaa2ff86fc719d5d0cb5050c14cbfbfa5b17-json.log",
"Name": "/addons-141392",
"RestartCount": 0,
"Driver": "overlay2",
"Platform": "linux",
"MountLabel": "",
"ProcessLabel": "",
"AppArmorProfile": "unconfined",
"ExecIDs": null,
"HostConfig": {
"Binds": [
"addons-141392:/var",
"/lib/modules:/lib/modules:ro"
],
"ContainerIDFile": "",
"LogConfig": {
"Type": "json-file",
"Config": {
"max-size": "100m"
}
},
"NetworkMode": "addons-141392",
"PortBindings": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": ""
}
]
},
"RestartPolicy": {
"Name": "no",
"MaximumRetryCount": 0
},
"AutoRemove": false,
"VolumeDriver": "",
"VolumesFrom": null,
"ConsoleSize": [
0,
0
],
"CapAdd": null,
"CapDrop": null,
"CgroupnsMode": "host",
"Dns": [],
"DnsOptions": [],
"DnsSearch": [],
"ExtraHosts": null,
"GroupAdd": null,
"IpcMode": "private",
"Cgroup": "",
"Links": null,
"OomScoreAdj": 0,
"PidMode": "",
"Privileged": true,
"PublishAllPorts": false,
"ReadonlyRootfs": false,
"SecurityOpt": [
"seccomp=unconfined",
"apparmor=unconfined",
"label=disable"
],
"Tmpfs": {
"/run": "",
"/tmp": ""
},
"UTSMode": "",
"UsernsMode": "",
"ShmSize": 67108864,
"Runtime": "runc",
"Isolation": "",
"CpuShares": 0,
"Memory": 4194304000,
"NanoCpus": 2000000000,
"CgroupParent": "",
"BlkioWeight": 0,
"BlkioWeightDevice": [],
"BlkioDeviceReadBps": [],
"BlkioDeviceWriteBps": [],
"BlkioDeviceReadIOps": [],
"BlkioDeviceWriteIOps": [],
"CpuPeriod": 0,
"CpuQuota": 0,
"CpuRealtimePeriod": 0,
"CpuRealtimeRuntime": 0,
"CpusetCpus": "",
"CpusetMems": "",
"Devices": [],
"DeviceCgroupRules": null,
"DeviceRequests": null,
"MemoryReservation": 0,
"MemorySwap": 8388608000,
"MemorySwappiness": null,
"OomKillDisable": false,
"PidsLimit": null,
"Ulimits": [],
"CpuCount": 0,
"CpuPercent": 0,
"IOMaximumIOps": 0,
"IOMaximumBandwidth": 0,
"MaskedPaths": null,
"ReadonlyPaths": null
},
"GraphDriver": {
"Data": {
"LowerDir": "/var/lib/docker/overlay2/b101d3d35792fbf3e832b6985a6115a62149a0c4eef4945f07242f2300bc0d21-init/diff:/var/lib/docker/overlay2/bb583947f58ae6325595139320a3fe1f2af4b5440b0a851b6250fc38726d88a3/diff",
"MergedDir": "/var/lib/docker/overlay2/b101d3d35792fbf3e832b6985a6115a62149a0c4eef4945f07242f2300bc0d21/merged",
"UpperDir": "/var/lib/docker/overlay2/b101d3d35792fbf3e832b6985a6115a62149a0c4eef4945f07242f2300bc0d21/diff",
"WorkDir": "/var/lib/docker/overlay2/b101d3d35792fbf3e832b6985a6115a62149a0c4eef4945f07242f2300bc0d21/work"
},
"Name": "overlay2"
},
"Mounts": [
{
"Type": "volume",
"Name": "addons-141392",
"Source": "/var/lib/docker/volumes/addons-141392/_data",
"Destination": "/var",
"Driver": "local",
"Mode": "z",
"RW": true,
"Propagation": ""
},
{
"Type": "bind",
"Source": "/lib/modules",
"Destination": "/lib/modules",
"Mode": "ro",
"RW": false,
"Propagation": "rprivate"
}
],
"Config": {
"Hostname": "addons-141392",
"Domainname": "",
"User": "",
"AttachStdin": false,
"AttachStdout": false,
"AttachStderr": false,
"ExposedPorts": {
"22/tcp": {},
"2376/tcp": {},
"32443/tcp": {},
"5000/tcp": {},
"8443/tcp": {}
},
"Tty": true,
"OpenStdin": false,
"StdinOnce": false,
"Env": [
"container=docker",
"PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
],
"Cmd": null,
"Image": "gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85",
"Volumes": null,
"WorkingDir": "/",
"Entrypoint": [
"/usr/local/bin/entrypoint",
"/sbin/init"
],
"OnBuild": null,
"Labels": {
"created_by.minikube.sigs.k8s.io": "true",
"mode.minikube.sigs.k8s.io": "addons-141392",
"name.minikube.sigs.k8s.io": "addons-141392",
"role.minikube.sigs.k8s.io": ""
},
"StopSignal": "SIGRTMIN+3"
},
"NetworkSettings": {
"Bridge": "",
"SandboxID": "ec2a6e0ea0708325e3afc3cc64e53a17800cb1d43c29c64b16f175100e44001b",
"SandboxKey": "/var/run/docker/netns/ec2a6e0ea070",
"Ports": {
"22/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32768"
}
],
"2376/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32769"
}
],
"32443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32772"
}
],
"5000/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32770"
}
],
"8443/tcp": [
{
"HostIp": "127.0.0.1",
"HostPort": "32771"
}
]
},
"HairpinMode": false,
"LinkLocalIPv6Address": "",
"LinkLocalIPv6PrefixLen": 0,
"SecondaryIPAddresses": null,
"SecondaryIPv6Addresses": null,
"EndpointID": "",
"Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"IPAddress": "",
"IPPrefixLen": 0,
"IPv6Gateway": "",
"MacAddress": "",
"Networks": {
"addons-141392": {
"IPAMConfig": {
"IPv4Address": "192.168.49.2"
},
"Links": null,
"Aliases": null,
"MacAddress": "02:42:c0:a8:31:02",
"DriverOpts": null,
"NetworkID": "70358e2ee875b17bd831d89f52f041dcfda73b1e26f7d9f0c066f0fa352a2192",
"EndpointID": "9aa04e82d331b50cac43fb3f112611032193d80e6b3a5866fd8a5c25a49628e6",
"Gateway": "192.168.49.1",
"IPAddress": "192.168.49.2",
"IPPrefixLen": 24,
"IPv6Gateway": "",
"GlobalIPv6Address": "",
"GlobalIPv6PrefixLen": 0,
"DNSNames": [
"addons-141392",
"f37408764470"
]
}
}
}
}
]
-- /stdout --
helpers_test.go:239: (dbg) Run: out/minikube-linux-amd64 status --format={{.Host}} -p addons-141392 -n addons-141392
helpers_test.go:244: <<< TestAddons/parallel/Registry FAILED: start of post-mortem logs <<<
helpers_test.go:245: ======> post-mortem[TestAddons/parallel/Registry]: minikube logs <======
helpers_test.go:247: (dbg) Run: out/minikube-linux-amd64 -p addons-141392 logs -n 25
helpers_test.go:252: TestAddons/parallel/Registry logs:
-- stdout --
==> Audit <==
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| Command | Args | Profile | User | Version | Start Time | End Time |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
| delete | -p download-docker-494880 | download-docker-494880 | jenkins | v1.34.0 | 06 Sep 24 18:29 UTC | 06 Sep 24 18:29 UTC |
| start | --download-only -p | binary-mirror-393189 | jenkins | v1.34.0 | 06 Sep 24 18:29 UTC | |
| | binary-mirror-393189 | | | | | |
| | --alsologtostderr | | | | | |
| | --binary-mirror | | | | | |
| | http://127.0.0.1:33265 | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| delete | -p binary-mirror-393189 | binary-mirror-393189 | jenkins | v1.34.0 | 06 Sep 24 18:29 UTC | 06 Sep 24 18:29 UTC |
| addons | enable dashboard -p | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:29 UTC | |
| | addons-141392 | | | | | |
| addons | disable dashboard -p | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:29 UTC | |
| | addons-141392 | | | | | |
| start | -p addons-141392 --wait=true | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:29 UTC | 06 Sep 24 18:32 UTC |
| | --memory=4000 --alsologtostderr | | | | | |
| | --addons=registry | | | | | |
| | --addons=metrics-server | | | | | |
| | --addons=volumesnapshots | | | | | |
| | --addons=csi-hostpath-driver | | | | | |
| | --addons=gcp-auth | | | | | |
| | --addons=cloud-spanner | | | | | |
| | --addons=inspektor-gadget | | | | | |
| | --addons=storage-provisioner-rancher | | | | | |
| | --addons=nvidia-device-plugin | | | | | |
| | --addons=yakd --addons=volcano | | | | | |
| | --driver=docker | | | | | |
| | --container-runtime=docker | | | | | |
| | --addons=ingress | | | | | |
| | --addons=ingress-dns | | | | | |
| | --addons=helm-tiller | | | | | |
| addons | addons-141392 addons disable | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:33 UTC | 06 Sep 24 18:33 UTC |
| | volcano --alsologtostderr -v=1 | | | | | |
| addons | disable cloud-spanner -p | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:41 UTC | 06 Sep 24 18:41 UTC |
| | addons-141392 | | | | | |
| ssh | addons-141392 ssh cat | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:41 UTC | 06 Sep 24 18:41 UTC |
| | /opt/local-path-provisioner/pvc-222bc931-c2ee-4dde-b88b-b8a6293d4be3_default_test-pvc/file1 | | | | | |
| addons | addons-141392 addons disable | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:41 UTC | 06 Sep 24 18:42 UTC |
| | storage-provisioner-rancher | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-141392 addons disable | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:41 UTC | 06 Sep 24 18:41 UTC |
| | yakd --alsologtostderr -v=1 | | | | | |
| addons | disable nvidia-device-plugin | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:41 UTC | 06 Sep 24 18:41 UTC |
| | -p addons-141392 | | | | | |
| addons | addons-141392 addons | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:41 UTC | 06 Sep 24 18:41 UTC |
| | disable metrics-server | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-141392 addons disable | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| | helm-tiller --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | enable headlamp | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| | -p addons-141392 | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-141392 addons | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| | disable csi-hostpath-driver | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | addons-141392 addons disable | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| | headlamp --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-141392 addons | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| | disable volumesnapshots | | | | | |
| | --alsologtostderr -v=1 | | | | | |
| addons | disable inspektor-gadget -p | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| | addons-141392 | | | | | |
| ssh | addons-141392 ssh curl -s | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| | http://127.0.0.1/ -H 'Host: | | | | | |
| | nginx.example.com' | | | | | |
| ip | addons-141392 ip | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| addons | addons-141392 addons disable | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| | ingress-dns --alsologtostderr | | | | | |
| | -v=1 | | | | | |
| addons | addons-141392 addons disable | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| | ingress --alsologtostderr -v=1 | | | | | |
| ip | addons-141392 ip | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| addons | addons-141392 addons disable | addons-141392 | jenkins | v1.34.0 | 06 Sep 24 18:42 UTC | 06 Sep 24 18:42 UTC |
| | registry --alsologtostderr | | | | | |
| | -v=1 | | | | | |
|---------|---------------------------------------------------------------------------------------------|------------------------|---------|---------|---------------------|---------------------|
==> Last Start <==
Log file created at: 2024/09/06 18:29:19
Running on machine: ubuntu-20-agent-14
Binary: Built with gc go1.22.5 for linux/amd64
Log line format: [IWEF]mmdd hh:mm:ss.uuuuuu threadid file:line] msg
I0906 18:29:19.287384 13912 out.go:345] Setting OutFile to fd 1 ...
I0906 18:29:19.287495 13912 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0906 18:29:19.287504 13912 out.go:358] Setting ErrFile to fd 2...
I0906 18:29:19.287510 13912 out.go:392] TERM=,COLORTERM=, which probably does not support color
I0906 18:29:19.287703 13912 root.go:338] Updating PATH: /home/jenkins/minikube-integration/19576-5848/.minikube/bin
I0906 18:29:19.288274 13912 out.go:352] Setting JSON to false
I0906 18:29:19.289064 13912 start.go:129] hostinfo: {"hostname":"ubuntu-20-agent-14","uptime":707,"bootTime":1725646652,"procs":174,"os":"linux","platform":"ubuntu","platformFamily":"debian","platformVersion":"20.04","kernelVersion":"5.15.0-1067-gcp","kernelArch":"x86_64","virtualizationSystem":"kvm","virtualizationRole":"guest","hostId":"591c9f12-2938-3743-e2bf-c56a050d43d1"}
I0906 18:29:19.289118 13912 start.go:139] virtualization: kvm guest
I0906 18:29:19.290888 13912 out.go:177] * [addons-141392] minikube v1.34.0 on Ubuntu 20.04 (kvm/amd64)
I0906 18:29:19.291941 13912 out.go:177] - MINIKUBE_LOCATION=19576
I0906 18:29:19.291965 13912 notify.go:220] Checking for updates...
I0906 18:29:19.293823 13912 out.go:177] - MINIKUBE_SUPPRESS_DOCKER_PERFORMANCE=true
I0906 18:29:19.294787 13912 out.go:177] - KUBECONFIG=/home/jenkins/minikube-integration/19576-5848/kubeconfig
I0906 18:29:19.295745 13912 out.go:177] - MINIKUBE_HOME=/home/jenkins/minikube-integration/19576-5848/.minikube
I0906 18:29:19.296630 13912 out.go:177] - MINIKUBE_BIN=out/minikube-linux-amd64
I0906 18:29:19.297565 13912 out.go:177] - MINIKUBE_FORCE_SYSTEMD=
I0906 18:29:19.298713 13912 driver.go:394] Setting default libvirt URI to qemu:///system
I0906 18:29:19.319007 13912 docker.go:123] docker version: linux-27.2.0:Docker Engine - Community
I0906 18:29:19.319119 13912 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0906 18:29:19.367610 13912 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-06 18:29:19.35906319 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1067-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86_
64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.0 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:472731909fa34bd7bc9c087e4c27943f9835f111 Expected:472731909fa34bd7bc9c087e4c27943f9835f111} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerErr
ors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0906 18:29:19.367715 13912 docker.go:318] overlay module found
I0906 18:29:19.369318 13912 out.go:177] * Using the docker driver based on user configuration
I0906 18:29:19.370486 13912 start.go:297] selected driver: docker
I0906 18:29:19.370498 13912 start.go:901] validating driver "docker" against <nil>
I0906 18:29:19.370508 13912 start.go:912] status for docker: {Installed:true Healthy:true Running:false NeedsImprovement:false Error:<nil> Reason: Fix: Doc: Version:}
I0906 18:29:19.371210 13912 cli_runner.go:164] Run: docker system info --format "{{json .}}"
I0906 18:29:19.414960 13912 info.go:266] docker info: {ID:TS6T:UINC:MIYS:RZPA:KS6T:4JQK:7JHN:D6RA:LDP2:MHAE:G32M:C5NQ Containers:0 ContainersRunning:0 ContainersPaused:0 ContainersStopped:0 Images:1 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Using metacopy false] [Native Overlay Diff true] [userxattr false]] SystemStatus:<nil> Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:<nil> Log:[awslogs fluentd gcplogs gelf journald json-file local splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:false KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6Tables:true Debug:false NFd:26 OomKillDisable:true NGoroutines:45 SystemTime:2024-09-06 18:29:19.407249789 +0000 UTC LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:5.15.0-1067-gcp OperatingSystem:Ubuntu 20.04.6 LTS OSType:linux Architecture:x86
_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:{AllowNondistributableArtifactsCIDRs:[] AllowNondistributableArtifactsHostnames:[] InsecureRegistryCIDRs:[127.0.0.0/8] IndexConfigs:{DockerIo:{Name:docker.io Mirrors:[] Secure:true Official:true}} Mirrors:[]} NCPU:8 MemTotal:33647939584 GenericResources:<nil> DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:ubuntu-20-agent-14 Labels:[] ExperimentalBuild:false ServerVersion:27.2.0 ClusterStore: ClusterAdvertise: Runtimes:{Runc:{Path:runc}} DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:<nil>} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:472731909fa34bd7bc9c087e4c27943f9835f111 Expected:472731909fa34bd7bc9c087e4c27943f9835f111} RuncCommit:{ID:v1.1.13-0-g58aa920 Expected:v1.1.13-0-g58aa920} InitCommit:{ID:de40ad0 Expected:de40ad0} SecurityOptions:[name=apparmor name=seccomp,profile=builtin] ProductLicense: Warnings:<nil> ServerEr
rors:[] ClientInfo:{Debug:false Plugins:[map[Name:buildx Path:/usr/libexec/docker/cli-plugins/docker-buildx SchemaVersion:0.1.0 ShortDescription:Docker Buildx Vendor:Docker Inc. Version:v0.16.2] map[Name:compose Path:/usr/libexec/docker/cli-plugins/docker-compose SchemaVersion:0.1.0 ShortDescription:Docker Compose Vendor:Docker Inc. Version:v2.29.2] map[Name:scan Path:/usr/libexec/docker/cli-plugins/docker-scan SchemaVersion:0.1.0 ShortDescription:Docker Scan Vendor:Docker Inc. Version:v0.23.0]] Warnings:<nil>}}
I0906 18:29:19.415122 13912 start_flags.go:310] no existing cluster config was found, will generate one from the flags
I0906 18:29:19.415326 13912 start_flags.go:947] Waiting for all components: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0906 18:29:19.416589 13912 out.go:177] * Using Docker driver with root privileges
I0906 18:29:19.417643 13912 cni.go:84] Creating CNI manager for ""
I0906 18:29:19.417663 13912 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0906 18:29:19.417673 13912 start_flags.go:319] Found "bridge CNI" CNI - setting NetworkPlugin=cni
I0906 18:29:19.417727 13912 start.go:340] cluster config:
{Name:addons-141392 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-141392 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: Netwo
rkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: Aut
oPauseInterval:1m0s}
I0906 18:29:19.418764 13912 out.go:177] * Starting "addons-141392" primary control-plane node in "addons-141392" cluster
I0906 18:29:19.419597 13912 cache.go:121] Beginning downloading kic base image for docker with docker
I0906 18:29:19.420475 13912 out.go:177] * Pulling base image v0.0.45 ...
I0906 18:29:19.421432 13912 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0906 18:29:19.421459 13912 preload.go:146] Found local preload: /home/jenkins/minikube-integration/19576-5848/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4
I0906 18:29:19.421456 13912 image.go:79] Checking for gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 in local docker daemon
I0906 18:29:19.421467 13912 cache.go:56] Caching tarball of preloaded images
I0906 18:29:19.421538 13912 preload.go:172] Found /home/jenkins/minikube-integration/19576-5848/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4 in cache, skipping download
I0906 18:29:19.421549 13912 cache.go:59] Finished verifying existence of preloaded tar for v1.31.0 on docker
I0906 18:29:19.421850 13912 profile.go:143] Saving config to /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/config.json ...
I0906 18:29:19.421871 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/config.json: {Name:mk2b1e42999a14a757fb27635c8059e4899e41c8 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:19.437195 13912 cache.go:149] Downloading gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 to local cache
I0906 18:29:19.437293 13912 image.go:63] Checking for gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 in local cache directory
I0906 18:29:19.437308 13912 image.go:66] Found gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 in local cache directory, skipping pull
I0906 18:29:19.437318 13912 image.go:135] gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 exists in cache, skipping pull
I0906 18:29:19.437325 13912 cache.go:152] successfully saved gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 as a tarball
I0906 18:29:19.437332 13912 cache.go:162] Loading gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 from local cache
I0906 18:29:31.113375 13912 cache.go:164] successfully loaded and using gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 from cached tarball
I0906 18:29:31.113415 13912 cache.go:194] Successfully downloaded all kic artifacts
I0906 18:29:31.113450 13912 start.go:360] acquireMachinesLock for addons-141392: {Name:mka88cae49e7fddc83e3e335b11587010ff3d6cb Clock:{} Delay:500ms Timeout:10m0s Cancel:<nil>}
I0906 18:29:31.113558 13912 start.go:364] duration metric: took 87.625µs to acquireMachinesLock for "addons-141392"
I0906 18:29:31.113584 13912 start.go:93] Provisioning new machine with config: &{Name:addons-141392 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-141392 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[
] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: Sock
etVMnetClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} &{Name: IP: Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0906 18:29:31.113687 13912 start.go:125] createHost starting for "" (driver="docker")
I0906 18:29:31.115232 13912 out.go:235] * Creating docker container (CPUs=2, Memory=4000MB) ...
I0906 18:29:31.115443 13912 start.go:159] libmachine.API.Create for "addons-141392" (driver="docker")
I0906 18:29:31.115479 13912 client.go:168] LocalClient.Create starting
I0906 18:29:31.115588 13912 main.go:141] libmachine: Creating CA: /home/jenkins/minikube-integration/19576-5848/.minikube/certs/ca.pem
I0906 18:29:31.287284 13912 main.go:141] libmachine: Creating client certificate: /home/jenkins/minikube-integration/19576-5848/.minikube/certs/cert.pem
I0906 18:29:31.422238 13912 cli_runner.go:164] Run: docker network inspect addons-141392 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
W0906 18:29:31.436996 13912 cli_runner.go:211] docker network inspect addons-141392 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}" returned with exit code 1
I0906 18:29:31.437047 13912 network_create.go:284] running [docker network inspect addons-141392] to gather additional debugging logs...
I0906 18:29:31.437063 13912 cli_runner.go:164] Run: docker network inspect addons-141392
W0906 18:29:31.451109 13912 cli_runner.go:211] docker network inspect addons-141392 returned with exit code 1
I0906 18:29:31.451132 13912 network_create.go:287] error running [docker network inspect addons-141392]: docker network inspect addons-141392: exit status 1
stdout:
[]
stderr:
Error response from daemon: network addons-141392 not found
I0906 18:29:31.451157 13912 network_create.go:289] output of [docker network inspect addons-141392]: -- stdout --
[]
-- /stdout --
** stderr **
Error response from daemon: network addons-141392 not found
** /stderr **
I0906 18:29:31.451576 13912 cli_runner.go:164] Run: docker network inspect bridge --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0906 18:29:31.466148 13912 network.go:206] using free private subnet 192.168.49.0/24: &{IP:192.168.49.0 Netmask:255.255.255.0 Prefix:24 CIDR:192.168.49.0/24 Gateway:192.168.49.1 ClientMin:192.168.49.2 ClientMax:192.168.49.254 Broadcast:192.168.49.255 IsPrivate:true Interface:{IfaceName: IfaceIPv4: IfaceMTU:0 IfaceMAC:} reservation:0xc001f8cff0}
I0906 18:29:31.466194 13912 network_create.go:124] attempt to create docker network addons-141392 192.168.49.0/24 with gateway 192.168.49.1 and MTU of 1500 ...
I0906 18:29:31.466238 13912 cli_runner.go:164] Run: docker network create --driver=bridge --subnet=192.168.49.0/24 --gateway=192.168.49.1 -o --ip-masq -o --icc -o com.docker.network.driver.mtu=1500 --label=created_by.minikube.sigs.k8s.io=true --label=name.minikube.sigs.k8s.io=addons-141392 addons-141392
I0906 18:29:31.524006 13912 network_create.go:108] docker network addons-141392 192.168.49.0/24 created
I0906 18:29:31.524035 13912 kic.go:121] calculated static IP "192.168.49.2" for the "addons-141392" container
I0906 18:29:31.524086 13912 cli_runner.go:164] Run: docker ps -a --format {{.Names}}
I0906 18:29:31.538225 13912 cli_runner.go:164] Run: docker volume create addons-141392 --label name.minikube.sigs.k8s.io=addons-141392 --label created_by.minikube.sigs.k8s.io=true
I0906 18:29:31.553348 13912 oci.go:103] Successfully created a docker volume addons-141392
I0906 18:29:31.553431 13912 cli_runner.go:164] Run: docker run --rm --name addons-141392-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-141392 --entrypoint /usr/bin/test -v addons-141392:/var gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 -d /var/lib
I0906 18:29:38.684018 13912 cli_runner.go:217] Completed: docker run --rm --name addons-141392-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-141392 --entrypoint /usr/bin/test -v addons-141392:/var gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 -d /var/lib: (7.130549937s)
I0906 18:29:38.684048 13912 oci.go:107] Successfully prepared a docker volume addons-141392
I0906 18:29:38.684071 13912 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0906 18:29:38.684090 13912 kic.go:194] Starting extracting preloaded images to volume ...
I0906 18:29:38.684163 13912 cli_runner.go:164] Run: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19576-5848/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-141392:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 -I lz4 -xf /preloaded.tar -C /extractDir
I0906 18:29:42.541063 13912 cli_runner.go:217] Completed: docker run --rm --entrypoint /usr/bin/tar -v /home/jenkins/minikube-integration/19576-5848/.minikube/cache/preloaded-tarball/preloaded-images-k8s-v18-v1.31.0-docker-overlay2-amd64.tar.lz4:/preloaded.tar:ro -v addons-141392:/extractDir gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 -I lz4 -xf /preloaded.tar -C /extractDir: (3.856862701s)
I0906 18:29:42.541091 13912 kic.go:203] duration metric: took 3.85699843s to extract preloaded images to volume ...
W0906 18:29:42.541218 13912 cgroups_linux.go:77] Your kernel does not support swap limit capabilities or the cgroup is not mounted.
I0906 18:29:42.541306 13912 cli_runner.go:164] Run: docker info --format "'{{json .SecurityOptions}}'"
I0906 18:29:42.585960 13912 cli_runner.go:164] Run: docker run -d -t --privileged --security-opt seccomp=unconfined --tmpfs /tmp --tmpfs /run -v /lib/modules:/lib/modules:ro --hostname addons-141392 --name addons-141392 --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=addons-141392 --label role.minikube.sigs.k8s.io= --label mode.minikube.sigs.k8s.io=addons-141392 --network addons-141392 --ip 192.168.49.2 --volume addons-141392:/var --security-opt apparmor=unconfined --memory=4000mb --cpus=2 -e container=docker --expose 8443 --publish=127.0.0.1::8443 --publish=127.0.0.1::22 --publish=127.0.0.1::2376 --publish=127.0.0.1::5000 --publish=127.0.0.1::32443 gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85
I0906 18:29:42.880122 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Running}}
I0906 18:29:42.895810 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:29:42.912454 13912 cli_runner.go:164] Run: docker exec addons-141392 stat /var/lib/dpkg/alternatives/iptables
I0906 18:29:42.951009 13912 oci.go:144] the created container "addons-141392" has a running status.
I0906 18:29:42.951041 13912 kic.go:225] Creating ssh key for kic: /home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa...
I0906 18:29:43.271466 13912 kic_runner.go:191] docker (temp): /home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa.pub --> /home/docker/.ssh/authorized_keys (381 bytes)
I0906 18:29:43.295138 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:29:43.312207 13912 kic_runner.go:93] Run: chown docker:docker /home/docker/.ssh/authorized_keys
I0906 18:29:43.312230 13912 kic_runner.go:114] Args: [docker exec --privileged addons-141392 chown docker:docker /home/docker/.ssh/authorized_keys]
I0906 18:29:43.379686 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:29:43.396474 13912 machine.go:93] provisionDockerMachine start ...
I0906 18:29:43.396583 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:43.413099 13912 main.go:141] libmachine: Using SSH client type: native
I0906 18:29:43.413350 13912 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0906 18:29:43.413370 13912 main.go:141] libmachine: About to run SSH command:
hostname
I0906 18:29:43.534711 13912 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-141392
I0906 18:29:43.534736 13912 ubuntu.go:169] provisioning hostname "addons-141392"
I0906 18:29:43.534782 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:43.551586 13912 main.go:141] libmachine: Using SSH client type: native
I0906 18:29:43.551753 13912 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0906 18:29:43.551765 13912 main.go:141] libmachine: About to run SSH command:
sudo hostname addons-141392 && echo "addons-141392" | sudo tee /etc/hostname
I0906 18:29:43.680408 13912 main.go:141] libmachine: SSH cmd err, output: <nil>: addons-141392
I0906 18:29:43.680479 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:43.696378 13912 main.go:141] libmachine: Using SSH client type: native
I0906 18:29:43.696656 13912 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0906 18:29:43.696689 13912 main.go:141] libmachine: About to run SSH command:
if ! grep -xq '.*\saddons-141392' /etc/hosts; then
if grep -xq '127.0.1.1\s.*' /etc/hosts; then
sudo sed -i 's/^127.0.1.1\s.*/127.0.1.1 addons-141392/g' /etc/hosts;
else
echo '127.0.1.1 addons-141392' | sudo tee -a /etc/hosts;
fi
fi
I0906 18:29:43.814715 13912 main.go:141] libmachine: SSH cmd err, output: <nil>:
I0906 18:29:43.814742 13912 ubuntu.go:175] set auth options {CertDir:/home/jenkins/minikube-integration/19576-5848/.minikube CaCertPath:/home/jenkins/minikube-integration/19576-5848/.minikube/certs/ca.pem CaPrivateKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/certs/ca-key.pem CaCertRemotePath:/etc/docker/ca.pem ServerCertPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/server.pem ServerKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/server-key.pem ClientKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/certs/key.pem ServerCertRemotePath:/etc/docker/server.pem ServerKeyRemotePath:/etc/docker/server-key.pem ClientCertPath:/home/jenkins/minikube-integration/19576-5848/.minikube/certs/cert.pem ServerCertSANs:[] StorePath:/home/jenkins/minikube-integration/19576-5848/.minikube}
I0906 18:29:43.814758 13912 ubuntu.go:177] setting up certificates
I0906 18:29:43.814768 13912 provision.go:84] configureAuth start
I0906 18:29:43.814814 13912 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-141392
I0906 18:29:43.831199 13912 provision.go:143] copyHostCerts
I0906 18:29:43.831269 13912 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19576-5848/.minikube/certs/ca.pem --> /home/jenkins/minikube-integration/19576-5848/.minikube/ca.pem (1082 bytes)
I0906 18:29:43.831384 13912 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19576-5848/.minikube/certs/cert.pem --> /home/jenkins/minikube-integration/19576-5848/.minikube/cert.pem (1123 bytes)
I0906 18:29:43.831461 13912 exec_runner.go:151] cp: /home/jenkins/minikube-integration/19576-5848/.minikube/certs/key.pem --> /home/jenkins/minikube-integration/19576-5848/.minikube/key.pem (1679 bytes)
I0906 18:29:43.831579 13912 provision.go:117] generating server cert: /home/jenkins/minikube-integration/19576-5848/.minikube/machines/server.pem ca-key=/home/jenkins/minikube-integration/19576-5848/.minikube/certs/ca.pem private-key=/home/jenkins/minikube-integration/19576-5848/.minikube/certs/ca-key.pem org=jenkins.addons-141392 san=[127.0.0.1 192.168.49.2 addons-141392 localhost minikube]
I0906 18:29:44.028295 13912 provision.go:177] copyRemoteCerts
I0906 18:29:44.028344 13912 ssh_runner.go:195] Run: sudo mkdir -p /etc/docker /etc/docker /etc/docker
I0906 18:29:44.028385 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:44.044189 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:29:44.131028 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/certs/ca.pem --> /etc/docker/ca.pem (1082 bytes)
I0906 18:29:44.150742 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/machines/server.pem --> /etc/docker/server.pem (1208 bytes)
I0906 18:29:44.169497 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/machines/server-key.pem --> /etc/docker/server-key.pem (1675 bytes)
I0906 18:29:44.188156 13912 provision.go:87] duration metric: took 373.379221ms to configureAuth
I0906 18:29:44.188175 13912 ubuntu.go:193] setting minikube options for container-runtime
I0906 18:29:44.188334 13912 config.go:182] Loaded profile config "addons-141392": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0906 18:29:44.188384 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:44.203424 13912 main.go:141] libmachine: Using SSH client type: native
I0906 18:29:44.203610 13912 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0906 18:29:44.203625 13912 main.go:141] libmachine: About to run SSH command:
df --output=fstype / | tail -n 1
I0906 18:29:44.323301 13912 main.go:141] libmachine: SSH cmd err, output: <nil>: overlay
I0906 18:29:44.323319 13912 ubuntu.go:71] root file system type: overlay
I0906 18:29:44.323429 13912 provision.go:314] Updating docker unit: /lib/systemd/system/docker.service ...
I0906 18:29:44.323478 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:44.339003 13912 main.go:141] libmachine: Using SSH client type: native
I0906 18:29:44.339163 13912 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0906 18:29:44.339222 13912 main.go:141] libmachine: About to run SSH command:
sudo mkdir -p /lib/systemd/system && printf %s "[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP \$MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
" | sudo tee /lib/systemd/system/docker.service.new
I0906 18:29:44.468036 13912 main.go:141] libmachine: SSH cmd err, output: <nil>: [Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
BindsTo=containerd.service
After=network-online.target firewalld.service containerd.service
Wants=network-online.target
Requires=docker.socket
StartLimitBurst=3
StartLimitIntervalSec=60
[Service]
Type=notify
Restart=on-failure
# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
# The base configuration already specifies an 'ExecStart=...' command. The first directive
# here is to clear out that command inherited from the base configuration. Without this,
# the command from the base configuration and the command specified here are treated as
# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
# will catch this invalid input and refuse to start the service with an error like:
# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
ExecStart=
ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
# Uncomment TasksMax if your systemd version supports it.
# Only systemd 226 and above support this version.
TasksMax=infinity
TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
[Install]
WantedBy=multi-user.target
I0906 18:29:44.468122 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:44.484102 13912 main.go:141] libmachine: Using SSH client type: native
I0906 18:29:44.484252 13912 main.go:141] libmachine: &{{{<nil> 0 [] [] []} docker [0x82f9c0] 0x832720 <nil> [] 0s} 127.0.0.1 32768 <nil> <nil>}
I0906 18:29:44.484268 13912 main.go:141] libmachine: About to run SSH command:
sudo diff -u /lib/systemd/system/docker.service /lib/systemd/system/docker.service.new || { sudo mv /lib/systemd/system/docker.service.new /lib/systemd/system/docker.service; sudo systemctl -f daemon-reload && sudo systemctl -f enable docker && sudo systemctl -f restart docker; }
I0906 18:29:45.110511 13912 main.go:141] libmachine: SSH cmd err, output: <nil>: --- /lib/systemd/system/docker.service 2024-08-27 14:13:43.000000000 +0000
+++ /lib/systemd/system/docker.service.new 2024-09-06 18:29:44.461754229 +0000
@@ -1,46 +1,49 @@
[Unit]
Description=Docker Application Container Engine
Documentation=https://docs.docker.com
-After=network-online.target docker.socket firewalld.service containerd.service time-set.target
-Wants=network-online.target containerd.service
+BindsTo=containerd.service
+After=network-online.target firewalld.service containerd.service
+Wants=network-online.target
Requires=docker.socket
+StartLimitBurst=3
+StartLimitIntervalSec=60
[Service]
Type=notify
-# the default is not to use systemd for cgroups because the delegate issues still
-# exists and systemd currently does not support the cgroup feature set required
-# for containers run by docker
-ExecStart=/usr/bin/dockerd -H fd:// --containerd=/run/containerd/containerd.sock
-ExecReload=/bin/kill -s HUP $MAINPID
-TimeoutStartSec=0
-RestartSec=2
-Restart=always
+Restart=on-failure
-# Note that StartLimit* options were moved from "Service" to "Unit" in systemd 229.
-# Both the old, and new location are accepted by systemd 229 and up, so using the old location
-# to make them work for either version of systemd.
-StartLimitBurst=3
-# Note that StartLimitInterval was renamed to StartLimitIntervalSec in systemd 230.
-# Both the old, and new name are accepted by systemd 230 and up, so using the old name to make
-# this option work for either version of systemd.
-StartLimitInterval=60s
+
+# This file is a systemd drop-in unit that inherits from the base dockerd configuration.
+# The base configuration already specifies an 'ExecStart=...' command. The first directive
+# here is to clear out that command inherited from the base configuration. Without this,
+# the command from the base configuration and the command specified here are treated as
+# a sequence of commands, which is not the desired behavior, nor is it valid -- systemd
+# will catch this invalid input and refuse to start the service with an error like:
+# Service has more than one ExecStart= setting, which is only allowed for Type=oneshot services.
+
+# NOTE: default-ulimit=nofile is set to an arbitrary number for consistency with other
+# container runtimes. If left unlimited, it may result in OOM issues with MySQL.
+ExecStart=
+ExecStart=/usr/bin/dockerd -H tcp://0.0.0.0:2376 -H unix:///var/run/docker.sock --default-ulimit=nofile=1048576:1048576 --tlsverify --tlscacert /etc/docker/ca.pem --tlscert /etc/docker/server.pem --tlskey /etc/docker/server-key.pem --label provider=docker --insecure-registry 10.96.0.0/12
+ExecReload=/bin/kill -s HUP $MAINPID
# Having non-zero Limit*s causes performance problems due to accounting overhead
# in the kernel. We recommend using cgroups to do container-local accounting.
+LimitNOFILE=infinity
LimitNPROC=infinity
LimitCORE=infinity
-# Comment TasksMax if your systemd version does not support it.
-# Only systemd 226 and above support this option.
+# Uncomment TasksMax if your systemd version supports it.
+# Only systemd 226 and above support this version.
TasksMax=infinity
+TimeoutStartSec=0
# set delegate yes so that systemd does not reset the cgroups of docker containers
Delegate=yes
# kill only the docker process, not all processes in the cgroup
KillMode=process
-OOMScoreAdjust=-500
[Install]
WantedBy=multi-user.target
Synchronizing state of docker.service with SysV service script with /lib/systemd/systemd-sysv-install.
Executing: /lib/systemd/systemd-sysv-install enable docker
I0906 18:29:45.110541 13912 machine.go:96] duration metric: took 1.71403253s to provisionDockerMachine
I0906 18:29:45.110558 13912 client.go:171] duration metric: took 13.995064009s to LocalClient.Create
I0906 18:29:45.110575 13912 start.go:167] duration metric: took 13.995132287s to libmachine.API.Create "addons-141392"
I0906 18:29:45.110584 13912 start.go:293] postStartSetup for "addons-141392" (driver="docker")
I0906 18:29:45.110595 13912 start.go:322] creating required directories: [/etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs]
I0906 18:29:45.110651 13912 ssh_runner.go:195] Run: sudo mkdir -p /etc/kubernetes/addons /etc/kubernetes/manifests /var/tmp/minikube /var/lib/minikube /var/lib/minikube/certs /var/lib/minikube/images /var/lib/minikube/binaries /tmp/gvisor /usr/share/ca-certificates /etc/ssl/certs
I0906 18:29:45.110683 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:45.125962 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:29:45.219383 13912 ssh_runner.go:195] Run: cat /etc/os-release
I0906 18:29:45.221973 13912 main.go:141] libmachine: Couldn't set key VERSION_CODENAME, no corresponding struct field found
I0906 18:29:45.222012 13912 main.go:141] libmachine: Couldn't set key PRIVACY_POLICY_URL, no corresponding struct field found
I0906 18:29:45.222025 13912 main.go:141] libmachine: Couldn't set key UBUNTU_CODENAME, no corresponding struct field found
I0906 18:29:45.222037 13912 info.go:137] Remote host: Ubuntu 22.04.4 LTS
I0906 18:29:45.222050 13912 filesync.go:126] Scanning /home/jenkins/minikube-integration/19576-5848/.minikube/addons for local assets ...
I0906 18:29:45.222101 13912 filesync.go:126] Scanning /home/jenkins/minikube-integration/19576-5848/.minikube/files for local assets ...
I0906 18:29:45.222130 13912 start.go:296] duration metric: took 111.539927ms for postStartSetup
I0906 18:29:45.222435 13912 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-141392
I0906 18:29:45.237280 13912 profile.go:143] Saving config to /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/config.json ...
I0906 18:29:45.237486 13912 ssh_runner.go:195] Run: sh -c "df -h /var | awk 'NR==2{print $5}'"
I0906 18:29:45.237519 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:45.252170 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:29:45.335363 13912 ssh_runner.go:195] Run: sh -c "df -BG /var | awk 'NR==2{print $4}'"
I0906 18:29:45.339036 13912 start.go:128] duration metric: took 14.225329333s to createHost
I0906 18:29:45.339058 13912 start.go:83] releasing machines lock for "addons-141392", held for 14.225486968s
I0906 18:29:45.339123 13912 cli_runner.go:164] Run: docker container inspect -f "{{range .NetworkSettings.Networks}}{{.IPAddress}},{{.GlobalIPv6Address}}{{end}}" addons-141392
I0906 18:29:45.354374 13912 ssh_runner.go:195] Run: cat /version.json
I0906 18:29:45.354423 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:45.354490 13912 ssh_runner.go:195] Run: curl -sS -m 2 https://registry.k8s.io/
I0906 18:29:45.354546 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:29:45.370845 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:29:45.371220 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:29:45.526025 13912 ssh_runner.go:195] Run: systemctl --version
I0906 18:29:45.529642 13912 ssh_runner.go:195] Run: sh -c "stat /etc/cni/net.d/*loopback.conf*"
I0906 18:29:45.533178 13912 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f -name *loopback.conf* -not -name *.mk_disabled -exec sh -c "grep -q loopback {} && ( grep -q name {} || sudo sed -i '/"type": "loopback"/i \ \ \ \ "name": "loopback",' {} ) && sudo sed -i 's|"cniVersion": ".*"|"cniVersion": "1.0.0"|g' {}" ;
I0906 18:29:45.553924 13912 cni.go:230] loopback cni configuration patched: "/etc/cni/net.d/*loopback.conf*" found
I0906 18:29:45.553972 13912 ssh_runner.go:195] Run: sudo find /etc/cni/net.d -maxdepth 1 -type f ( ( -name *bridge* -or -name *podman* ) -and -not -name *.mk_disabled ) -printf "%p, " -exec sh -c "sudo mv {} {}.mk_disabled" ;
I0906 18:29:45.576057 13912 cni.go:262] disabled [/etc/cni/net.d/87-podman-bridge.conflist, /etc/cni/net.d/100-crio-bridge.conf] bridge cni config(s)
I0906 18:29:45.576076 13912 start.go:495] detecting cgroup driver to use...
I0906 18:29:45.576104 13912 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0906 18:29:45.576208 13912 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///run/containerd/containerd.sock
" | sudo tee /etc/crictl.yaml"
I0906 18:29:45.588821 13912 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)sandbox_image = .*$|\1sandbox_image = "registry.k8s.io/pause:3.10"|' /etc/containerd/config.toml"
I0906 18:29:45.596427 13912 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)restrict_oom_score_adj = .*$|\1restrict_oom_score_adj = false|' /etc/containerd/config.toml"
I0906 18:29:45.604116 13912 containerd.go:146] configuring containerd to use "cgroupfs" as cgroup driver...
I0906 18:29:45.604161 13912 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)SystemdCgroup = .*$|\1SystemdCgroup = false|g' /etc/containerd/config.toml"
I0906 18:29:45.611544 13912 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runtime.v1.linux"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0906 18:29:45.618896 13912 ssh_runner.go:195] Run: sh -c "sudo sed -i '/systemd_cgroup/d' /etc/containerd/config.toml"
I0906 18:29:45.626135 13912 ssh_runner.go:195] Run: sh -c "sudo sed -i 's|"io.containerd.runc.v1"|"io.containerd.runc.v2"|g' /etc/containerd/config.toml"
I0906 18:29:45.633538 13912 ssh_runner.go:195] Run: sh -c "sudo rm -rf /etc/cni/net.mk"
I0906 18:29:45.640615 13912 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)conf_dir = .*$|\1conf_dir = "/etc/cni/net.d"|g' /etc/containerd/config.toml"
I0906 18:29:45.648320 13912 ssh_runner.go:195] Run: sh -c "sudo sed -i '/^ *enable_unprivileged_ports = .*/d' /etc/containerd/config.toml"
I0906 18:29:45.655672 13912 ssh_runner.go:195] Run: sh -c "sudo sed -i -r 's|^( *)\[plugins."io.containerd.grpc.v1.cri"\]|&\n\1 enable_unprivileged_ports = true|' /etc/containerd/config.toml"
I0906 18:29:45.663124 13912 ssh_runner.go:195] Run: sudo sysctl net.bridge.bridge-nf-call-iptables
I0906 18:29:45.669549 13912 ssh_runner.go:195] Run: sudo sh -c "echo 1 > /proc/sys/net/ipv4/ip_forward"
I0906 18:29:45.675948 13912 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0906 18:29:45.747232 13912 ssh_runner.go:195] Run: sudo systemctl restart containerd
I0906 18:29:45.814860 13912 start.go:495] detecting cgroup driver to use...
I0906 18:29:45.814903 13912 detect.go:187] detected "cgroupfs" cgroup driver on host os
I0906 18:29:45.814946 13912 ssh_runner.go:195] Run: sudo systemctl cat docker.service
I0906 18:29:45.824887 13912 cruntime.go:279] skipping containerd shutdown because we are bound to it
I0906 18:29:45.824945 13912 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service crio
I0906 18:29:45.834427 13912 ssh_runner.go:195] Run: /bin/bash -c "sudo mkdir -p /etc && printf %s "runtime-endpoint: unix:///var/run/cri-dockerd.sock
" | sudo tee /etc/crictl.yaml"
I0906 18:29:45.848299 13912 ssh_runner.go:195] Run: which cri-dockerd
I0906 18:29:45.851375 13912 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/cri-docker.service.d
I0906 18:29:45.858944 13912 ssh_runner.go:362] scp memory --> /etc/systemd/system/cri-docker.service.d/10-cni.conf (190 bytes)
I0906 18:29:45.873617 13912 ssh_runner.go:195] Run: sudo systemctl unmask docker.service
I0906 18:29:45.948391 13912 ssh_runner.go:195] Run: sudo systemctl enable docker.socket
I0906 18:29:46.037784 13912 docker.go:574] configuring docker to use "cgroupfs" as cgroup driver...
I0906 18:29:46.037923 13912 ssh_runner.go:362] scp memory --> /etc/docker/daemon.json (130 bytes)
I0906 18:29:46.053087 13912 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0906 18:29:46.131651 13912 ssh_runner.go:195] Run: sudo systemctl restart docker
I0906 18:29:46.357996 13912 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.socket
I0906 18:29:46.367982 13912 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0906 18:29:46.377935 13912 ssh_runner.go:195] Run: sudo systemctl unmask cri-docker.socket
I0906 18:29:46.461566 13912 ssh_runner.go:195] Run: sudo systemctl enable cri-docker.socket
I0906 18:29:46.532428 13912 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0906 18:29:46.600732 13912 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.socket
I0906 18:29:46.611669 13912 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service cri-docker.service
I0906 18:29:46.620572 13912 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0906 18:29:46.696634 13912 ssh_runner.go:195] Run: sudo systemctl restart cri-docker.service
I0906 18:29:46.752093 13912 start.go:542] Will wait 60s for socket path /var/run/cri-dockerd.sock
I0906 18:29:46.752171 13912 ssh_runner.go:195] Run: stat /var/run/cri-dockerd.sock
I0906 18:29:46.755249 13912 start.go:563] Will wait 60s for crictl version
I0906 18:29:46.755299 13912 ssh_runner.go:195] Run: which crictl
I0906 18:29:46.758071 13912 ssh_runner.go:195] Run: sudo /usr/bin/crictl version
I0906 18:29:46.787029 13912 start.go:579] Version: 0.1.0
RuntimeName: docker
RuntimeVersion: 27.2.0
RuntimeApiVersion: v1
I0906 18:29:46.787081 13912 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0906 18:29:46.807561 13912 ssh_runner.go:195] Run: docker version --format {{.Server.Version}}
I0906 18:29:46.830522 13912 out.go:235] * Preparing Kubernetes v1.31.0 on Docker 27.2.0 ...
I0906 18:29:46.830608 13912 cli_runner.go:164] Run: docker network inspect addons-141392 --format "{"Name": "{{.Name}}","Driver": "{{.Driver}}","Subnet": "{{range .IPAM.Config}}{{.Subnet}}{{end}}","Gateway": "{{range .IPAM.Config}}{{.Gateway}}{{end}}","MTU": {{if (index .Options "com.docker.network.driver.mtu")}}{{(index .Options "com.docker.network.driver.mtu")}}{{else}}0{{end}}, "ContainerIPs": [{{range $k,$v := .Containers }}"{{$v.IPv4Address}}",{{end}}]}"
I0906 18:29:46.846324 13912 ssh_runner.go:195] Run: grep 192.168.49.1 host.minikube.internal$ /etc/hosts
I0906 18:29:46.849423 13912 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\thost.minikube.internal$' "/etc/hosts"; echo "192.168.49.1 host.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0906 18:29:46.858765 13912 kubeadm.go:883] updating cluster {Name:addons-141392 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-141392 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] D
NSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnet
ClientPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s} ...
I0906 18:29:46.858866 13912 preload.go:131] Checking if preload exists for k8s version v1.31.0 and runtime docker
I0906 18:29:46.858905 13912 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0906 18:29:46.876199 13912 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-apiserver:v1.31.0
registry.k8s.io/kube-controller-manager:v1.31.0
registry.k8s.io/kube-scheduler:v1.31.0
registry.k8s.io/kube-proxy:v1.31.0
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
registry.k8s.io/coredns/coredns:v1.11.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0906 18:29:46.876216 13912 docker.go:615] Images already preloaded, skipping extraction
I0906 18:29:46.876250 13912 ssh_runner.go:195] Run: docker images --format {{.Repository}}:{{.Tag}}
I0906 18:29:46.892286 13912 docker.go:685] Got preloaded images: -- stdout --
registry.k8s.io/kube-controller-manager:v1.31.0
registry.k8s.io/kube-scheduler:v1.31.0
registry.k8s.io/kube-apiserver:v1.31.0
registry.k8s.io/kube-proxy:v1.31.0
registry.k8s.io/etcd:3.5.15-0
registry.k8s.io/pause:3.10
registry.k8s.io/coredns/coredns:v1.11.1
gcr.io/k8s-minikube/storage-provisioner:v5
-- /stdout --
I0906 18:29:46.892302 13912 cache_images.go:84] Images are preloaded, skipping loading
I0906 18:29:46.892317 13912 kubeadm.go:934] updating node { 192.168.49.2 8443 v1.31.0 docker true true} ...
I0906 18:29:46.892401 13912 kubeadm.go:946] kubelet [Unit]
Wants=docker.socket
[Service]
ExecStart=
ExecStart=/var/lib/minikube/binaries/v1.31.0/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --config=/var/lib/kubelet/config.yaml --hostname-override=addons-141392 --kubeconfig=/etc/kubernetes/kubelet.conf --node-ip=192.168.49.2
[Install]
config:
{KubernetesVersion:v1.31.0 ClusterName:addons-141392 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSDomain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:}
I0906 18:29:46.892444 13912 ssh_runner.go:195] Run: docker info --format {{.CgroupDriver}}
I0906 18:29:46.933571 13912 cni.go:84] Creating CNI manager for ""
I0906 18:29:46.933611 13912 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0906 18:29:46.933636 13912 kubeadm.go:84] Using pod CIDR: 10.244.0.0/16
I0906 18:29:46.933661 13912 kubeadm.go:181] kubeadm options: {CertDir:/var/lib/minikube/certs ServiceCIDR:10.96.0.0/12 PodSubnet:10.244.0.0/16 AdvertiseAddress:192.168.49.2 APIServerPort:8443 KubernetesVersion:v1.31.0 EtcdDataDir:/var/lib/minikube/etcd EtcdExtraArgs:map[] ClusterName:addons-141392 NodeName:addons-141392 DNSDomain:cluster.local CRISocket:/var/run/cri-dockerd.sock ImageRepository: ComponentOptions:[{Component:apiServer ExtraArgs:map[enable-admission-plugins:NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota] Pairs:map[certSANs:["127.0.0.1", "localhost", "192.168.49.2"]]} {Component:controllerManager ExtraArgs:map[allocate-node-cidrs:true leader-elect:false] Pairs:map[]} {Component:scheduler ExtraArgs:map[leader-elect:false] Pairs:map[]}] FeatureArgs:map[] NodeIP:192.168.49.2 CgroupDriver:cgroupfs ClientCAFile:/var/lib/minikube/certs/ca.crt StaticPodPath:/etc/kuber
netes/manifests ControlPlaneAddress:control-plane.minikube.internal KubeProxyOptions:map[] ResolvConfSearchRegression:false KubeletConfigOpts:map[containerRuntimeEndpoint:unix:///var/run/cri-dockerd.sock hairpinMode:hairpin-veth runtimeRequestTimeout:15m] PrependCriSocketUnix:true}
I0906 18:29:46.933833 13912 kubeadm.go:187] kubeadm config:
apiVersion: kubeadm.k8s.io/v1beta3
kind: InitConfiguration
localAPIEndpoint:
advertiseAddress: 192.168.49.2
bindPort: 8443
bootstrapTokens:
- groups:
- system:bootstrappers:kubeadm:default-node-token
ttl: 24h0m0s
usages:
- signing
- authentication
nodeRegistration:
criSocket: unix:///var/run/cri-dockerd.sock
name: "addons-141392"
kubeletExtraArgs:
node-ip: 192.168.49.2
taints: []
---
apiVersion: kubeadm.k8s.io/v1beta3
kind: ClusterConfiguration
apiServer:
certSANs: ["127.0.0.1", "localhost", "192.168.49.2"]
extraArgs:
enable-admission-plugins: "NamespaceLifecycle,LimitRanger,ServiceAccount,DefaultStorageClass,DefaultTolerationSeconds,NodeRestriction,MutatingAdmissionWebhook,ValidatingAdmissionWebhook,ResourceQuota"
controllerManager:
extraArgs:
allocate-node-cidrs: "true"
leader-elect: "false"
scheduler:
extraArgs:
leader-elect: "false"
certificatesDir: /var/lib/minikube/certs
clusterName: mk
controlPlaneEndpoint: control-plane.minikube.internal:8443
etcd:
local:
dataDir: /var/lib/minikube/etcd
extraArgs:
proxy-refresh-interval: "70000"
kubernetesVersion: v1.31.0
networking:
dnsDomain: cluster.local
podSubnet: "10.244.0.0/16"
serviceSubnet: 10.96.0.0/12
---
apiVersion: kubelet.config.k8s.io/v1beta1
kind: KubeletConfiguration
authentication:
x509:
clientCAFile: /var/lib/minikube/certs/ca.crt
cgroupDriver: cgroupfs
containerRuntimeEndpoint: unix:///var/run/cri-dockerd.sock
hairpinMode: hairpin-veth
runtimeRequestTimeout: 15m
clusterDomain: "cluster.local"
# disable disk resource management by default
imageGCHighThresholdPercent: 100
evictionHard:
nodefs.available: "0%"
nodefs.inodesFree: "0%"
imagefs.available: "0%"
failSwapOn: false
staticPodPath: /etc/kubernetes/manifests
---
apiVersion: kubeproxy.config.k8s.io/v1alpha1
kind: KubeProxyConfiguration
clusterCIDR: "10.244.0.0/16"
metricsBindAddress: 0.0.0.0:10249
conntrack:
maxPerCore: 0
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_established"
tcpEstablishedTimeout: 0s
# Skip setting "net.netfilter.nf_conntrack_tcp_timeout_close"
tcpCloseWaitTimeout: 0s
I0906 18:29:46.933892 13912 ssh_runner.go:195] Run: sudo ls /var/lib/minikube/binaries/v1.31.0
I0906 18:29:46.941304 13912 binaries.go:44] Found k8s binaries, skipping transfer
I0906 18:29:46.941356 13912 ssh_runner.go:195] Run: sudo mkdir -p /etc/systemd/system/kubelet.service.d /lib/systemd/system /var/tmp/minikube
I0906 18:29:46.948259 13912 ssh_runner.go:362] scp memory --> /etc/systemd/system/kubelet.service.d/10-kubeadm.conf (312 bytes)
I0906 18:29:46.962567 13912 ssh_runner.go:362] scp memory --> /lib/systemd/system/kubelet.service (352 bytes)
I0906 18:29:46.976608 13912 ssh_runner.go:362] scp memory --> /var/tmp/minikube/kubeadm.yaml.new (2155 bytes)
I0906 18:29:46.990813 13912 ssh_runner.go:195] Run: grep 192.168.49.2 control-plane.minikube.internal$ /etc/hosts
I0906 18:29:46.993584 13912 ssh_runner.go:195] Run: /bin/bash -c "{ grep -v $'\tcontrol-plane.minikube.internal$' "/etc/hosts"; echo "192.168.49.2 control-plane.minikube.internal"; } > /tmp/h.$$; sudo cp /tmp/h.$$ "/etc/hosts""
I0906 18:29:47.002339 13912 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0906 18:29:47.076953 13912 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0906 18:29:47.088315 13912 certs.go:68] Setting up /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392 for IP: 192.168.49.2
I0906 18:29:47.088336 13912 certs.go:194] generating shared ca certs ...
I0906 18:29:47.088356 13912 certs.go:226] acquiring lock for ca certs: {Name:mk7b9e4ebaf47d18ef837233f5d1697cd2643544 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.088487 13912 certs.go:240] generating "minikubeCA" ca cert: /home/jenkins/minikube-integration/19576-5848/.minikube/ca.key
I0906 18:29:47.348352 13912 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19576-5848/.minikube/ca.crt ...
I0906 18:29:47.348378 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/ca.crt: {Name:mk2b91cfd978e40685543cc955259cef5cc8c84e Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.348532 13912 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19576-5848/.minikube/ca.key ...
I0906 18:29:47.348541 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/ca.key: {Name:mkcdd119b7d2d27d4c03b777d30f0d5e3b88b367 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.348615 13912 certs.go:240] generating "proxyClientCA" ca cert: /home/jenkins/minikube-integration/19576-5848/.minikube/proxy-client-ca.key
I0906 18:29:47.513756 13912 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19576-5848/.minikube/proxy-client-ca.crt ...
I0906 18:29:47.513787 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/proxy-client-ca.crt: {Name:mkfa6b9e7bfd919bd32eed817a5cc287cc649939 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.513972 13912 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19576-5848/.minikube/proxy-client-ca.key ...
I0906 18:29:47.513984 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/proxy-client-ca.key: {Name:mk8c7e0011f45a97174d6b5647b8648cb1ad59a2 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.514062 13912 certs.go:256] generating profile certs ...
I0906 18:29:47.514120 13912 certs.go:363] generating signed profile cert for "minikube-user": /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/client.key
I0906 18:29:47.514135 13912 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/client.crt with IP's: []
I0906 18:29:47.567117 13912 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/client.crt ...
I0906 18:29:47.567146 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/client.crt: {Name:mk469fb8ac495a3af3fa0a65c26b73c4c5a82201 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.567297 13912 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/client.key ...
I0906 18:29:47.567308 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/client.key: {Name:mk76768249c03938c291a1676a4e4977e7eee4db Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.567381 13912 certs.go:363] generating signed profile cert for "minikube": /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.key.71068b69
I0906 18:29:47.567398 13912 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.crt.71068b69 with IP's: [10.96.0.1 127.0.0.1 10.0.0.1 192.168.49.2]
I0906 18:29:47.726855 13912 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.crt.71068b69 ...
I0906 18:29:47.726881 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.crt.71068b69: {Name:mk81a82a4e678767ffa950f4b32b056220859870 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.727017 13912 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.key.71068b69 ...
I0906 18:29:47.727030 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.key.71068b69: {Name:mkbf965015f17ed85623cc343a71bc6abc6e4e41 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.727097 13912 certs.go:381] copying /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.crt.71068b69 -> /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.crt
I0906 18:29:47.727162 13912 certs.go:385] copying /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.key.71068b69 -> /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.key
I0906 18:29:47.727206 13912 certs.go:363] generating signed profile cert for "aggregator": /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/proxy-client.key
I0906 18:29:47.727221 13912 crypto.go:68] Generating cert /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/proxy-client.crt with IP's: []
I0906 18:29:47.830550 13912 crypto.go:156] Writing cert to /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/proxy-client.crt ...
I0906 18:29:47.830574 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/proxy-client.crt: {Name:mk1bf549274ba98ee00720561e55ff0e8b42da6d Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.830715 13912 crypto.go:164] Writing key to /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/proxy-client.key ...
I0906 18:29:47.830725 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/proxy-client.key: {Name:mk9655398e21089dfc6724f1dc477a718c4dbc2a Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:29:47.830881 13912 certs.go:484] found cert: /home/jenkins/minikube-integration/19576-5848/.minikube/certs/ca-key.pem (1679 bytes)
I0906 18:29:47.830910 13912 certs.go:484] found cert: /home/jenkins/minikube-integration/19576-5848/.minikube/certs/ca.pem (1082 bytes)
I0906 18:29:47.830934 13912 certs.go:484] found cert: /home/jenkins/minikube-integration/19576-5848/.minikube/certs/cert.pem (1123 bytes)
I0906 18:29:47.830957 13912 certs.go:484] found cert: /home/jenkins/minikube-integration/19576-5848/.minikube/certs/key.pem (1679 bytes)
I0906 18:29:47.831505 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/ca.crt --> /var/lib/minikube/certs/ca.crt (1111 bytes)
I0906 18:29:47.851802 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/ca.key --> /var/lib/minikube/certs/ca.key (1675 bytes)
I0906 18:29:47.871018 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/proxy-client-ca.crt --> /var/lib/minikube/certs/proxy-client-ca.crt (1119 bytes)
I0906 18:29:47.890248 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/proxy-client-ca.key --> /var/lib/minikube/certs/proxy-client-ca.key (1675 bytes)
I0906 18:29:47.909188 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.crt --> /var/lib/minikube/certs/apiserver.crt (1419 bytes)
I0906 18:29:47.928176 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/apiserver.key --> /var/lib/minikube/certs/apiserver.key (1675 bytes)
I0906 18:29:47.946932 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/proxy-client.crt --> /var/lib/minikube/certs/proxy-client.crt (1147 bytes)
I0906 18:29:47.965857 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/profiles/addons-141392/proxy-client.key --> /var/lib/minikube/certs/proxy-client.key (1675 bytes)
I0906 18:29:47.985756 13912 ssh_runner.go:362] scp /home/jenkins/minikube-integration/19576-5848/.minikube/ca.crt --> /usr/share/ca-certificates/minikubeCA.pem (1111 bytes)
I0906 18:29:48.005411 13912 ssh_runner.go:362] scp memory --> /var/lib/minikube/kubeconfig (738 bytes)
I0906 18:29:48.019475 13912 ssh_runner.go:195] Run: openssl version
I0906 18:29:48.023790 13912 ssh_runner.go:195] Run: sudo /bin/bash -c "test -s /usr/share/ca-certificates/minikubeCA.pem && ln -fs /usr/share/ca-certificates/minikubeCA.pem /etc/ssl/certs/minikubeCA.pem"
I0906 18:29:48.031115 13912 ssh_runner.go:195] Run: ls -la /usr/share/ca-certificates/minikubeCA.pem
I0906 18:29:48.033825 13912 certs.go:528] hashing: -rw-r--r-- 1 root root 1111 Sep 6 18:29 /usr/share/ca-certificates/minikubeCA.pem
I0906 18:29:48.033864 13912 ssh_runner.go:195] Run: openssl x509 -hash -noout -in /usr/share/ca-certificates/minikubeCA.pem
I0906 18:29:48.039533 13912 ssh_runner.go:195] Run: sudo /bin/bash -c "test -L /etc/ssl/certs/b5213941.0 || ln -fs /etc/ssl/certs/minikubeCA.pem /etc/ssl/certs/b5213941.0"
I0906 18:29:48.046659 13912 ssh_runner.go:195] Run: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt
I0906 18:29:48.049234 13912 certs.go:399] 'apiserver-kubelet-client' cert doesn't exist, likely first start: stat /var/lib/minikube/certs/apiserver-kubelet-client.crt: Process exited with status 1
stdout:
stderr:
stat: cannot statx '/var/lib/minikube/certs/apiserver-kubelet-client.crt': No such file or directory
I0906 18:29:48.049282 13912 kubeadm.go:392] StartCluster: {Name:addons-141392 KeepContext:false EmbedCerts:false MinikubeISO: KicBaseImage:gcr.io/k8s-minikube/kicbase:v0.0.45@sha256:81df288595202a317b1a4dc2506ca2e4ed5f22373c19a441b88cfbf4b9867c85 Memory:4000 CPUs:2 DiskSize:20000 Driver:docker HyperkitVpnKitSock: HyperkitVSockPorts:[] DockerEnv:[] ContainerVolumeMounts:[] InsecureRegistry:[] RegistryMirror:[] HostOnlyCIDR:192.168.59.1/24 HypervVirtualSwitch: HypervUseExternalSwitch:false HypervExternalAdapter: KVMNetwork:default KVMQemuURI:qemu:///system KVMGPU:false KVMHidden:false KVMNUMACount:1 APIServerPort:8443 DockerOpt:[] DisableDriverMounts:false NFSShare:[] NFSSharesRoot:/nfsshares UUID: NoVTXCheck:false DNSProxy:false HostDNSResolver:true HostOnlyNicType:virtio NatNicType:virtio SSHIPAddress: SSHUser:root SSHKey: SSHPort:22 KubernetesConfig:{KubernetesVersion:v1.31.0 ClusterName:addons-141392 Namespace:default APIServerHAVIP: APIServerName:minikubeCA APIServerNames:[] APIServerIPs:[] DNSD
omain:cluster.local ContainerRuntime:docker CRISocket: NetworkPlugin:cni FeatureGates: ServiceCIDR:10.96.0.0/12 ImageRepository: LoadBalancerStartIP: LoadBalancerEndIP: CustomIngressCert: RegistryAliases: ExtraOptions:[] ShouldLoadCachedImages:true EnableDefaultCNI:false CNI:} Nodes:[{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}] Addons:map[] CustomAddonImages:map[] CustomAddonRegistries:map[] VerifyComponents:map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true] StartHostTimeout:6m0s ScheduledStop:<nil> ExposedPorts:[] ListenAddress: Network: Subnet: MultiNodeRequested:false ExtraDisks:0 CertExpiration:26280h0m0s Mount:false MountString:/home/jenkins:/minikube-host Mount9PVersion:9p2000.L MountGID:docker MountIP: MountMSize:262144 MountOptions:[] MountPort:0 MountType:9p MountUID:docker BinaryMirror: DisableOptimizations:false DisableMetrics:false CustomQemuFirmwarePath: SocketVMnetCli
entPath: SocketVMnetPath: StaticIP: SSHAuthSock: SSHAgentPID:0 GPUs: AutoPauseInterval:1m0s}
I0906 18:29:48.049393 13912 ssh_runner.go:195] Run: docker ps --filter status=paused --filter=name=k8s_.*_(kube-system)_ --format={{.ID}}
I0906 18:29:48.065070 13912 ssh_runner.go:195] Run: sudo ls /var/lib/kubelet/kubeadm-flags.env /var/lib/kubelet/config.yaml /var/lib/minikube/etcd
I0906 18:29:48.072033 13912 ssh_runner.go:195] Run: sudo cp /var/tmp/minikube/kubeadm.yaml.new /var/tmp/minikube/kubeadm.yaml
I0906 18:29:48.078729 13912 kubeadm.go:214] ignoring SystemVerification for kubeadm because of docker driver
I0906 18:29:48.078763 13912 ssh_runner.go:195] Run: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf
I0906 18:29:48.085394 13912 kubeadm.go:155] config check failed, skipping stale config cleanup: sudo ls -la /etc/kubernetes/admin.conf /etc/kubernetes/kubelet.conf /etc/kubernetes/controller-manager.conf /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
ls: cannot access '/etc/kubernetes/admin.conf': No such file or directory
ls: cannot access '/etc/kubernetes/kubelet.conf': No such file or directory
ls: cannot access '/etc/kubernetes/controller-manager.conf': No such file or directory
ls: cannot access '/etc/kubernetes/scheduler.conf': No such file or directory
I0906 18:29:48.085407 13912 kubeadm.go:157] found existing configuration files:
I0906 18:29:48.085433 13912 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf
I0906 18:29:48.092197 13912 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/admin.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/admin.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/admin.conf: No such file or directory
I0906 18:29:48.092226 13912 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/admin.conf
I0906 18:29:48.098803 13912 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf
I0906 18:29:48.105491 13912 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/kubelet.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/kubelet.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/kubelet.conf: No such file or directory
I0906 18:29:48.105530 13912 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/kubelet.conf
I0906 18:29:48.112029 13912 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf
I0906 18:29:48.118630 13912 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/controller-manager.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/controller-manager.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/controller-manager.conf: No such file or directory
I0906 18:29:48.118664 13912 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/controller-manager.conf
I0906 18:29:48.125234 13912 ssh_runner.go:195] Run: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf
I0906 18:29:48.132161 13912 kubeadm.go:163] "https://control-plane.minikube.internal:8443" may not be in /etc/kubernetes/scheduler.conf - will remove: sudo grep https://control-plane.minikube.internal:8443 /etc/kubernetes/scheduler.conf: Process exited with status 2
stdout:
stderr:
grep: /etc/kubernetes/scheduler.conf: No such file or directory
I0906 18:29:48.132193 13912 ssh_runner.go:195] Run: sudo rm -f /etc/kubernetes/scheduler.conf
I0906 18:29:48.138587 13912 ssh_runner.go:286] Start: /bin/bash -c "sudo env PATH="/var/lib/minikube/binaries/v1.31.0:$PATH" kubeadm init --config /var/tmp/minikube/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests,DirAvailable--var-lib-minikube,DirAvailable--var-lib-minikube-etcd,FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml,FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml,FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml,FileAvailable--etc-kubernetes-manifests-etcd.yaml,Port-10250,Swap,NumCPU,Mem,SystemVerification,FileContent--proc-sys-net-bridge-bridge-nf-call-iptables"
I0906 18:29:48.167548 13912 kubeadm.go:310] W0906 18:29:48.167004 1922 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "ClusterConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0906 18:29:48.168174 13912 kubeadm.go:310] W0906 18:29:48.167571 1922 common.go:101] your configuration file uses a deprecated API spec: "kubeadm.k8s.io/v1beta3" (kind: "InitConfiguration"). Please use 'kubeadm config migrate --old-config old.yaml --new-config new.yaml', which will write the new, similar spec using a newer API version.
I0906 18:29:48.186505 13912 kubeadm.go:310] [WARNING SystemVerification]: failed to parse kernel config: unable to load kernel module: "configs", output: "modprobe: FATAL: Module configs not found in directory /lib/modules/5.15.0-1067-gcp\n", err: exit status 1
I0906 18:29:48.231777 13912 kubeadm.go:310] [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service'
I0906 18:29:56.823465 13912 kubeadm.go:310] [init] Using Kubernetes version: v1.31.0
I0906 18:29:56.823549 13912 kubeadm.go:310] [preflight] Running pre-flight checks
I0906 18:29:56.823662 13912 kubeadm.go:310] [preflight] The system verification failed. Printing the output from the verification:
I0906 18:29:56.823738 13912 kubeadm.go:310] [0;37mKERNEL_VERSION[0m: [0;32m5.15.0-1067-gcp[0m
I0906 18:29:56.823777 13912 kubeadm.go:310] [0;37mOS[0m: [0;32mLinux[0m
I0906 18:29:56.823838 13912 kubeadm.go:310] [0;37mCGROUPS_CPU[0m: [0;32menabled[0m
I0906 18:29:56.823910 13912 kubeadm.go:310] [0;37mCGROUPS_CPUACCT[0m: [0;32menabled[0m
I0906 18:29:56.823991 13912 kubeadm.go:310] [0;37mCGROUPS_CPUSET[0m: [0;32menabled[0m
I0906 18:29:56.824074 13912 kubeadm.go:310] [0;37mCGROUPS_DEVICES[0m: [0;32menabled[0m
I0906 18:29:56.824154 13912 kubeadm.go:310] [0;37mCGROUPS_FREEZER[0m: [0;32menabled[0m
I0906 18:29:56.824228 13912 kubeadm.go:310] [0;37mCGROUPS_MEMORY[0m: [0;32menabled[0m
I0906 18:29:56.824296 13912 kubeadm.go:310] [0;37mCGROUPS_PIDS[0m: [0;32menabled[0m
I0906 18:29:56.824370 13912 kubeadm.go:310] [0;37mCGROUPS_HUGETLB[0m: [0;32menabled[0m
I0906 18:29:56.824431 13912 kubeadm.go:310] [0;37mCGROUPS_BLKIO[0m: [0;32menabled[0m
I0906 18:29:56.824527 13912 kubeadm.go:310] [preflight] Pulling images required for setting up a Kubernetes cluster
I0906 18:29:56.824644 13912 kubeadm.go:310] [preflight] This might take a minute or two, depending on the speed of your internet connection
I0906 18:29:56.824749 13912 kubeadm.go:310] [preflight] You can also perform this action beforehand using 'kubeadm config images pull'
I0906 18:29:56.824822 13912 kubeadm.go:310] [certs] Using certificateDir folder "/var/lib/minikube/certs"
I0906 18:29:56.826145 13912 out.go:235] - Generating certificates and keys ...
I0906 18:29:56.826206 13912 kubeadm.go:310] [certs] Using existing ca certificate authority
I0906 18:29:56.826259 13912 kubeadm.go:310] [certs] Using existing apiserver certificate and key on disk
I0906 18:29:56.826318 13912 kubeadm.go:310] [certs] Generating "apiserver-kubelet-client" certificate and key
I0906 18:29:56.826381 13912 kubeadm.go:310] [certs] Generating "front-proxy-ca" certificate and key
I0906 18:29:56.826449 13912 kubeadm.go:310] [certs] Generating "front-proxy-client" certificate and key
I0906 18:29:56.826517 13912 kubeadm.go:310] [certs] Generating "etcd/ca" certificate and key
I0906 18:29:56.826596 13912 kubeadm.go:310] [certs] Generating "etcd/server" certificate and key
I0906 18:29:56.826769 13912 kubeadm.go:310] [certs] etcd/server serving cert is signed for DNS names [addons-141392 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0906 18:29:56.826841 13912 kubeadm.go:310] [certs] Generating "etcd/peer" certificate and key
I0906 18:29:56.826970 13912 kubeadm.go:310] [certs] etcd/peer serving cert is signed for DNS names [addons-141392 localhost] and IPs [192.168.49.2 127.0.0.1 ::1]
I0906 18:29:56.827030 13912 kubeadm.go:310] [certs] Generating "etcd/healthcheck-client" certificate and key
I0906 18:29:56.827089 13912 kubeadm.go:310] [certs] Generating "apiserver-etcd-client" certificate and key
I0906 18:29:56.827126 13912 kubeadm.go:310] [certs] Generating "sa" key and public key
I0906 18:29:56.827179 13912 kubeadm.go:310] [kubeconfig] Using kubeconfig folder "/etc/kubernetes"
I0906 18:29:56.827226 13912 kubeadm.go:310] [kubeconfig] Writing "admin.conf" kubeconfig file
I0906 18:29:56.827277 13912 kubeadm.go:310] [kubeconfig] Writing "super-admin.conf" kubeconfig file
I0906 18:29:56.827331 13912 kubeadm.go:310] [kubeconfig] Writing "kubelet.conf" kubeconfig file
I0906 18:29:56.827385 13912 kubeadm.go:310] [kubeconfig] Writing "controller-manager.conf" kubeconfig file
I0906 18:29:56.827435 13912 kubeadm.go:310] [kubeconfig] Writing "scheduler.conf" kubeconfig file
I0906 18:29:56.827505 13912 kubeadm.go:310] [etcd] Creating static Pod manifest for local etcd in "/etc/kubernetes/manifests"
I0906 18:29:56.827589 13912 kubeadm.go:310] [control-plane] Using manifest folder "/etc/kubernetes/manifests"
I0906 18:29:56.829715 13912 out.go:235] - Booting up control plane ...
I0906 18:29:56.829783 13912 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-apiserver"
I0906 18:29:56.829849 13912 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-controller-manager"
I0906 18:29:56.829920 13912 kubeadm.go:310] [control-plane] Creating static Pod manifest for "kube-scheduler"
I0906 18:29:56.830030 13912 kubeadm.go:310] [kubelet-start] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env"
I0906 18:29:56.830132 13912 kubeadm.go:310] [kubelet-start] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml"
I0906 18:29:56.830194 13912 kubeadm.go:310] [kubelet-start] Starting the kubelet
I0906 18:29:56.830330 13912 kubeadm.go:310] [wait-control-plane] Waiting for the kubelet to boot up the control plane as static Pods from directory "/etc/kubernetes/manifests"
I0906 18:29:56.830424 13912 kubeadm.go:310] [kubelet-check] Waiting for a healthy kubelet at http://127.0.0.1:10248/healthz. This can take up to 4m0s
I0906 18:29:56.830473 13912 kubeadm.go:310] [kubelet-check] The kubelet is healthy after 501.431151ms
I0906 18:29:56.830536 13912 kubeadm.go:310] [api-check] Waiting for a healthy API server. This can take up to 4m0s
I0906 18:29:56.830585 13912 kubeadm.go:310] [api-check] The API server is healthy after 4.501590458s
I0906 18:29:56.830675 13912 kubeadm.go:310] [upload-config] Storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace
I0906 18:29:56.830785 13912 kubeadm.go:310] [kubelet] Creating a ConfigMap "kubelet-config" in namespace kube-system with the configuration for the kubelets in the cluster
I0906 18:29:56.830834 13912 kubeadm.go:310] [upload-certs] Skipping phase. Please see --upload-certs
I0906 18:29:56.830979 13912 kubeadm.go:310] [mark-control-plane] Marking the node addons-141392 as control-plane by adding the labels: [node-role.kubernetes.io/control-plane node.kubernetes.io/exclude-from-external-load-balancers]
I0906 18:29:56.831026 13912 kubeadm.go:310] [bootstrap-token] Using token: ctu886.v16oj0700kc8ckj7
I0906 18:29:56.832086 13912 out.go:235] - Configuring RBAC rules ...
I0906 18:29:56.832165 13912 kubeadm.go:310] [bootstrap-token] Configuring bootstrap tokens, cluster-info ConfigMap, RBAC Roles
I0906 18:29:56.832236 13912 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to get nodes
I0906 18:29:56.832370 13912 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials
I0906 18:29:56.832489 13912 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token
I0906 18:29:56.832583 13912 kubeadm.go:310] [bootstrap-token] Configured RBAC rules to allow certificate rotation for all node client certificates in the cluster
I0906 18:29:56.832657 13912 kubeadm.go:310] [bootstrap-token] Creating the "cluster-info" ConfigMap in the "kube-public" namespace
I0906 18:29:56.832750 13912 kubeadm.go:310] [kubelet-finalize] Updating "/etc/kubernetes/kubelet.conf" to point to a rotatable kubelet client certificate and key
I0906 18:29:56.832788 13912 kubeadm.go:310] [addons] Applied essential addon: CoreDNS
I0906 18:29:56.832854 13912 kubeadm.go:310] [addons] Applied essential addon: kube-proxy
I0906 18:29:56.832867 13912 kubeadm.go:310]
I0906 18:29:56.832951 13912 kubeadm.go:310] Your Kubernetes control-plane has initialized successfully!
I0906 18:29:56.832964 13912 kubeadm.go:310]
I0906 18:29:56.833068 13912 kubeadm.go:310] To start using your cluster, you need to run the following as a regular user:
I0906 18:29:56.833075 13912 kubeadm.go:310]
I0906 18:29:56.833096 13912 kubeadm.go:310] mkdir -p $HOME/.kube
I0906 18:29:56.833146 13912 kubeadm.go:310] sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
I0906 18:29:56.833188 13912 kubeadm.go:310] sudo chown $(id -u):$(id -g) $HOME/.kube/config
I0906 18:29:56.833194 13912 kubeadm.go:310]
I0906 18:29:56.833239 13912 kubeadm.go:310] Alternatively, if you are the root user, you can run:
I0906 18:29:56.833245 13912 kubeadm.go:310]
I0906 18:29:56.833310 13912 kubeadm.go:310] export KUBECONFIG=/etc/kubernetes/admin.conf
I0906 18:29:56.833320 13912 kubeadm.go:310]
I0906 18:29:56.833399 13912 kubeadm.go:310] You should now deploy a pod network to the cluster.
I0906 18:29:56.833493 13912 kubeadm.go:310] Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
I0906 18:29:56.833564 13912 kubeadm.go:310] https://kubernetes.io/docs/concepts/cluster-administration/addons/
I0906 18:29:56.833571 13912 kubeadm.go:310]
I0906 18:29:56.833640 13912 kubeadm.go:310] You can now join any number of control-plane nodes by copying certificate authorities
I0906 18:29:56.833703 13912 kubeadm.go:310] and service account keys on each node and then running the following as root:
I0906 18:29:56.833709 13912 kubeadm.go:310]
I0906 18:29:56.833786 13912 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token ctu886.v16oj0700kc8ckj7 \
I0906 18:29:56.833876 13912 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:397571e80128d906650d787cf9171ede7802a762ec148b7ac750f5c625bb97d3 \
I0906 18:29:56.833895 13912 kubeadm.go:310] --control-plane
I0906 18:29:56.833901 13912 kubeadm.go:310]
I0906 18:29:56.833967 13912 kubeadm.go:310] Then you can join any number of worker nodes by running the following on each as root:
I0906 18:29:56.833973 13912 kubeadm.go:310]
I0906 18:29:56.834038 13912 kubeadm.go:310] kubeadm join control-plane.minikube.internal:8443 --token ctu886.v16oj0700kc8ckj7 \
I0906 18:29:56.834152 13912 kubeadm.go:310] --discovery-token-ca-cert-hash sha256:397571e80128d906650d787cf9171ede7802a762ec148b7ac750f5c625bb97d3
I0906 18:29:56.834169 13912 cni.go:84] Creating CNI manager for ""
I0906 18:29:56.834189 13912 cni.go:158] "docker" driver + "docker" container runtime found on kubernetes v1.24+, recommending bridge
I0906 18:29:56.835502 13912 out.go:177] * Configuring bridge CNI (Container Networking Interface) ...
I0906 18:29:56.836649 13912 ssh_runner.go:195] Run: sudo mkdir -p /etc/cni/net.d
I0906 18:29:56.844704 13912 ssh_runner.go:362] scp memory --> /etc/cni/net.d/1-k8s.conflist (496 bytes)
I0906 18:29:56.859807 13912 ssh_runner.go:195] Run: /bin/bash -c "cat /proc/$(pgrep kube-apiserver)/oom_adj"
I0906 18:29:56.859860 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl create clusterrolebinding minikube-rbac --clusterrole=cluster-admin --serviceaccount=kube-system:default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:29:56.859883 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig label --overwrite nodes addons-141392 minikube.k8s.io/updated_at=2024_09_06T18_29_56_0700 minikube.k8s.io/version=v1.34.0 minikube.k8s.io/commit=e6b6435971a63e36b5096cd544634422129cef13 minikube.k8s.io/name=addons-141392 minikube.k8s.io/primary=true
I0906 18:29:56.866593 13912 ops.go:34] apiserver oom_adj: -16
I0906 18:29:56.925353 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:29:57.426043 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:29:57.926226 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:29:58.425550 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:29:58.926107 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:29:59.425704 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:29:59.926177 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:30:00.426233 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:30:00.925633 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:30:01.425492 13912 ssh_runner.go:195] Run: sudo /var/lib/minikube/binaries/v1.31.0/kubectl get sa default --kubeconfig=/var/lib/minikube/kubeconfig
I0906 18:30:01.487129 13912 kubeadm.go:1113] duration metric: took 4.627310852s to wait for elevateKubeSystemPrivileges
I0906 18:30:01.487162 13912 kubeadm.go:394] duration metric: took 13.437883004s to StartCluster
I0906 18:30:01.487183 13912 settings.go:142] acquiring lock: {Name:mk91579767b9ce139f79002b1be861bd63e61c75 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:30:01.487300 13912 settings.go:150] Updating kubeconfig: /home/jenkins/minikube-integration/19576-5848/kubeconfig
I0906 18:30:01.487788 13912 lock.go:35] WriteFile acquiring /home/jenkins/minikube-integration/19576-5848/kubeconfig: {Name:mka96677f1fb48e42f1f7a69452e140337533c37 Clock:{} Delay:500ms Timeout:1m0s Cancel:<nil>}
I0906 18:30:01.487973 13912 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml"
I0906 18:30:01.488020 13912 start.go:235] Will wait 6m0s for node &{Name: IP:192.168.49.2 Port:8443 KubernetesVersion:v1.31.0 ContainerRuntime:docker ControlPlane:true Worker:true}
I0906 18:30:01.488074 13912 addons.go:507] enable addons start: toEnable=map[ambassador:false auto-pause:false cloud-spanner:true csi-hostpath-driver:true dashboard:false default-storageclass:true efk:false freshpod:false gcp-auth:true gvisor:false headlamp:false helm-tiller:true inaccel:false ingress:true ingress-dns:true inspektor-gadget:true istio:false istio-provisioner:false kong:false kubeflow:false kubevirt:false logviewer:false metallb:false metrics-server:true nvidia-device-plugin:true nvidia-driver-installer:false nvidia-gpu-device-plugin:false olm:false pod-security-policy:false portainer:false registry:true registry-aliases:false registry-creds:false storage-provisioner:true storage-provisioner-gluster:false storage-provisioner-rancher:true volcano:true volumesnapshots:true yakd:true]
I0906 18:30:01.488186 13912 addons.go:69] Setting yakd=true in profile "addons-141392"
I0906 18:30:01.488217 13912 addons.go:234] Setting addon yakd=true in "addons-141392"
I0906 18:30:01.488216 13912 addons.go:69] Setting inspektor-gadget=true in profile "addons-141392"
I0906 18:30:01.488250 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.488253 13912 config.go:182] Loaded profile config "addons-141392": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0906 18:30:01.488273 13912 addons.go:69] Setting volumesnapshots=true in profile "addons-141392"
I0906 18:30:01.488284 13912 addons.go:234] Setting addon inspektor-gadget=true in "addons-141392"
I0906 18:30:01.488277 13912 addons.go:69] Setting storage-provisioner=true in profile "addons-141392"
I0906 18:30:01.488308 13912 addons.go:234] Setting addon volumesnapshots=true in "addons-141392"
I0906 18:30:01.488337 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.488656 13912 addons.go:69] Setting csi-hostpath-driver=true in profile "addons-141392"
I0906 18:30:01.488767 13912 addons.go:234] Setting addon csi-hostpath-driver=true in "addons-141392"
I0906 18:30:01.488808 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.488330 13912 addons.go:69] Setting cloud-spanner=true in profile "addons-141392"
I0906 18:30:01.488832 13912 addons.go:69] Setting storage-provisioner-rancher=true in profile "addons-141392"
I0906 18:30:01.488312 13912 addons.go:69] Setting gcp-auth=true in profile "addons-141392"
I0906 18:30:01.488885 13912 addons_storage_classes.go:33] enableOrDisableStorageClasses storage-provisioner-rancher=true on "addons-141392"
I0906 18:30:01.488900 13912 addons.go:69] Setting ingress=true in profile "addons-141392"
I0906 18:30:01.488339 13912 addons.go:234] Setting addon storage-provisioner=true in "addons-141392"
I0906 18:30:01.488754 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.488952 13912 addons.go:69] Setting ingress-dns=true in profile "addons-141392"
I0906 18:30:01.488982 13912 addons.go:234] Setting addon ingress-dns=true in "addons-141392"
I0906 18:30:01.489021 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.489035 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.489242 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.489457 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.489561 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.489571 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.489581 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.488618 13912 addons.go:69] Setting volcano=true in profile "addons-141392"
I0906 18:30:01.489861 13912 addons.go:234] Setting addon volcano=true in "addons-141392"
I0906 18:30:01.489896 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.488908 13912 addons.go:69] Setting default-storageclass=true in profile "addons-141392"
I0906 18:30:01.490148 13912 addons_storage_classes.go:33] enableOrDisableStorageClasses default-storageclass=true on "addons-141392"
I0906 18:30:01.490412 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.490484 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.490796 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.490986 13912 addons.go:69] Setting metrics-server=true in profile "addons-141392"
I0906 18:30:01.491022 13912 addons.go:234] Setting addon metrics-server=true in "addons-141392"
I0906 18:30:01.491057 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.489027 13912 addons.go:234] Setting addon cloud-spanner=true in "addons-141392"
I0906 18:30:01.491184 13912 addons.go:69] Setting nvidia-device-plugin=true in profile "addons-141392"
I0906 18:30:01.492054 13912 addons.go:234] Setting addon nvidia-device-plugin=true in "addons-141392"
I0906 18:30:01.492066 13912 addons.go:69] Setting registry=true in profile "addons-141392"
I0906 18:30:01.492102 13912 addons.go:234] Setting addon registry=true in "addons-141392"
I0906 18:30:01.492104 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.492137 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.492152 13912 addons.go:69] Setting helm-tiller=true in profile "addons-141392"
I0906 18:30:01.492187 13912 addons.go:234] Setting addon helm-tiller=true in "addons-141392"
I0906 18:30:01.492212 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.492667 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.492714 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.492751 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.494653 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.495051 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.496271 13912 out.go:177] * Verifying Kubernetes components...
I0906 18:30:01.497520 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.497077 13912 mustload.go:65] Loading cluster: addons-141392
I0906 18:30:01.498000 13912 config.go:182] Loaded profile config "addons-141392": Driver=docker, ContainerRuntime=docker, KubernetesVersion=v1.31.0
I0906 18:30:01.498310 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.499308 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.500292 13912 ssh_runner.go:195] Run: sudo systemctl daemon-reload
I0906 18:30:01.488926 13912 addons.go:234] Setting addon ingress=true in "addons-141392"
I0906 18:30:01.500969 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.501460 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.535595 13912 out.go:177] - Using image docker.io/marcnuri/yakd:0.0.5
I0906 18:30:01.536966 13912 addons.go:431] installing /etc/kubernetes/addons/yakd-ns.yaml
I0906 18:30:01.537020 13912 ssh_runner.go:362] scp yakd/yakd-ns.yaml --> /etc/kubernetes/addons/yakd-ns.yaml (171 bytes)
I0906 18:30:01.537211 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.544421 13912 out.go:177] - Using image gcr.io/k8s-minikube/minikube-ingress-dns:0.0.3
I0906 18:30:01.545679 13912 addons.go:431] installing /etc/kubernetes/addons/ingress-dns-pod.yaml
I0906 18:30:01.545699 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-dns-pod.yaml (2442 bytes)
I0906 18:30:01.545759 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.559786 13912 out.go:177] - Using image docker.io/volcanosh/vc-scheduler:v1.9.0
I0906 18:30:01.559794 13912 out.go:177] - Using image ghcr.io/inspektor-gadget/inspektor-gadget:v0.32.0
I0906 18:30:01.560084 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.560117 13912 out.go:177] - Using image registry.k8s.io/sig-storage/snapshot-controller:v6.1.0
I0906 18:30:01.561908 13912 out.go:177] - Using image docker.io/volcanosh/vc-webhook-manager:v1.9.0
I0906 18:30:01.561972 13912 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0906 18:30:01.562846 13912 addons.go:234] Setting addon storage-provisioner-rancher=true in "addons-141392"
I0906 18:30:01.562885 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.563340 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.564287 13912 out.go:177] - Using image docker.io/volcanosh/vc-controller-manager:v1.9.0
I0906 18:30:01.564340 13912 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0906 18:30:01.564387 13912 addons.go:431] installing /etc/kubernetes/addons/ig-namespace.yaml
I0906 18:30:01.564398 13912 ssh_runner.go:362] scp inspektor-gadget/ig-namespace.yaml --> /etc/kubernetes/addons/ig-namespace.yaml (55 bytes)
I0906 18:30:01.564454 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.564747 13912 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml
I0906 18:30:01.564759 13912 ssh_runner.go:362] scp volumesnapshots/csi-hostpath-snapshotclass.yaml --> /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml (934 bytes)
I0906 18:30:01.564816 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.568703 13912 addons.go:431] installing /etc/kubernetes/addons/volcano-deployment.yaml
I0906 18:30:01.568728 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volcano-deployment.yaml (434001 bytes)
I0906 18:30:01.568781 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.573377 13912 out.go:177] - Using image registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.6.0
I0906 18:30:01.574210 13912 out.go:177] - Using image registry.k8s.io/ingress-nginx/controller:v1.11.2
I0906 18:30:01.574480 13912 addons.go:234] Setting addon default-storageclass=true in "addons-141392"
I0906 18:30:01.574699 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:01.575249 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:01.576654 13912 addons.go:431] installing /etc/kubernetes/addons/ingress-deploy.yaml
I0906 18:30:01.576681 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ingress-deploy.yaml (16078 bytes)
I0906 18:30:01.576742 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.577057 13912 out.go:177] - Using image registry.k8s.io/sig-storage/hostpathplugin:v1.9.0
I0906 18:30:01.578144 13912 out.go:177] - Using image gcr.io/k8s-minikube/storage-provisioner:v5
I0906 18:30:01.578297 13912 out.go:177] - Using image nvcr.io/nvidia/k8s-device-plugin:v0.16.2
I0906 18:30:01.579353 13912 out.go:177] - Using image registry.k8s.io/sig-storage/livenessprobe:v2.8.0
I0906 18:30:01.579502 13912 addons.go:431] installing /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0906 18:30:01.579529 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/nvidia-device-plugin.yaml (1966 bytes)
I0906 18:30:01.579582 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.579805 13912 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner.yaml
I0906 18:30:01.579816 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner.yaml (2676 bytes)
I0906 18:30:01.579859 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.588782 13912 out.go:177] - Using image registry.k8s.io/sig-storage/csi-resizer:v1.6.0
I0906 18:30:01.589936 13912 out.go:177] - Using image registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0
I0906 18:30:01.590943 13912 out.go:177] - Using image registry.k8s.io/sig-storage/csi-provisioner:v3.3.0
I0906 18:30:01.593259 13912 out.go:177] - Using image registry.k8s.io/sig-storage/csi-attacher:v4.0.0
I0906 18:30:01.594480 13912 out.go:177] - Using image docker.io/registry:2.8.3
I0906 18:30:01.594506 13912 out.go:177] - Using image registry.k8s.io/sig-storage/csi-external-health-monitor-controller:v0.7.0
I0906 18:30:01.595612 13912 addons.go:431] installing /etc/kubernetes/addons/rbac-external-attacher.yaml
I0906 18:30:01.595632 13912 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-attacher.yaml --> /etc/kubernetes/addons/rbac-external-attacher.yaml (3073 bytes)
I0906 18:30:01.595705 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.598097 13912 out.go:177] - Using image gcr.io/k8s-minikube/kube-registry-proxy:0.0.6
I0906 18:30:01.599479 13912 addons.go:431] installing /etc/kubernetes/addons/registry-rc.yaml
I0906 18:30:01.599497 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-rc.yaml (860 bytes)
I0906 18:30:01.599616 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.625123 13912 out.go:177] - Using image gcr.io/cloud-spanner-emulator/emulator:1.5.23
I0906 18:30:01.626364 13912 addons.go:431] installing /etc/kubernetes/addons/deployment.yaml
I0906 18:30:01.626386 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/deployment.yaml (1004 bytes)
I0906 18:30:01.626450 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.633365 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.636341 13912 out.go:177] - Using image docker.io/rancher/local-path-provisioner:v0.0.22
I0906 18:30:01.637577 13912 out.go:177] - Using image docker.io/busybox:stable
I0906 18:30:01.637691 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.638266 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.638895 13912 out.go:177] - Using image ghcr.io/helm/tiller:v2.17.0
I0906 18:30:01.639040 13912 addons.go:431] installing /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0906 18:30:01.639059 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/storage-provisioner-rancher.yaml (3113 bytes)
I0906 18:30:01.639108 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.638898 13912 out.go:177] - Using image registry.k8s.io/metrics-server/metrics-server:v0.7.2
I0906 18:30:01.640662 13912 addons.go:431] installing /etc/kubernetes/addons/metrics-apiservice.yaml
I0906 18:30:01.640680 13912 ssh_runner.go:362] scp metrics-server/metrics-apiservice.yaml --> /etc/kubernetes/addons/metrics-apiservice.yaml (424 bytes)
I0906 18:30:01.640733 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.642724 13912 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-dp.yaml
I0906 18:30:01.642745 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/helm-tiller-dp.yaml (2422 bytes)
I0906 18:30:01.642797 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.652804 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.655609 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.661121 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.663756 13912 addons.go:431] installing /etc/kubernetes/addons/storageclass.yaml
I0906 18:30:01.663775 13912 ssh_runner.go:362] scp storageclass/storageclass.yaml --> /etc/kubernetes/addons/storageclass.yaml (271 bytes)
I0906 18:30:01.663833 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:01.663747 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.666532 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.667023 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.669223 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.669490 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.675608 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.686535 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
W0906 18:30:01.690210 13912 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0906 18:30:01.690242 13912 retry.go:31] will retry after 330.476069ms: ssh: handshake failed: EOF
W0906 18:30:01.690335 13912 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0906 18:30:01.690344 13912 retry.go:31] will retry after 135.420394ms: ssh: handshake failed: EOF
I0906 18:30:01.698461 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:01.700693 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
W0906 18:30:01.704037 13912 sshutil.go:64] dial failure (will retry): ssh: handshake failed: EOF
I0906 18:30:01.704062 13912 retry.go:31] will retry after 187.938695ms: ssh: handshake failed: EOF
I0906 18:30:01.897729 13912 ssh_runner.go:195] Run: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -"
I0906 18:30:01.897873 13912 ssh_runner.go:195] Run: sudo systemctl start kubelet
I0906 18:30:02.077050 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml
I0906 18:30:02.086284 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml
I0906 18:30:02.087246 13912 addons.go:431] installing /etc/kubernetes/addons/yakd-sa.yaml
I0906 18:30:02.087269 13912 ssh_runner.go:362] scp yakd/yakd-sa.yaml --> /etc/kubernetes/addons/yakd-sa.yaml (247 bytes)
I0906 18:30:02.179633 13912 addons.go:431] installing /etc/kubernetes/addons/rbac-hostpath.yaml
I0906 18:30:02.179711 13912 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-hostpath.yaml --> /etc/kubernetes/addons/rbac-hostpath.yaml (4266 bytes)
I0906 18:30:02.192442 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml
I0906 18:30:02.193335 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml
I0906 18:30:02.272742 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml
I0906 18:30:02.281267 13912 addons.go:431] installing /etc/kubernetes/addons/yakd-crb.yaml
I0906 18:30:02.281346 13912 ssh_runner.go:362] scp yakd/yakd-crb.yaml --> /etc/kubernetes/addons/yakd-crb.yaml (422 bytes)
I0906 18:30:02.284837 13912 addons.go:431] installing /etc/kubernetes/addons/registry-svc.yaml
I0906 18:30:02.284864 13912 ssh_runner.go:362] scp registry/registry-svc.yaml --> /etc/kubernetes/addons/registry-svc.yaml (398 bytes)
I0906 18:30:02.292193 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml
I0906 18:30:02.373274 13912 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-rbac.yaml
I0906 18:30:02.373299 13912 ssh_runner.go:362] scp helm-tiller/helm-tiller-rbac.yaml --> /etc/kubernetes/addons/helm-tiller-rbac.yaml (1188 bytes)
I0906 18:30:02.377887 13912 addons.go:431] installing /etc/kubernetes/addons/metrics-server-deployment.yaml
I0906 18:30:02.377910 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/metrics-server-deployment.yaml (1907 bytes)
I0906 18:30:02.381076 13912 addons.go:431] installing /etc/kubernetes/addons/ig-serviceaccount.yaml
I0906 18:30:02.381103 13912 ssh_runner.go:362] scp inspektor-gadget/ig-serviceaccount.yaml --> /etc/kubernetes/addons/ig-serviceaccount.yaml (80 bytes)
I0906 18:30:02.387556 13912 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml
I0906 18:30:02.387590 13912 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotclasses.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml (6471 bytes)
I0906 18:30:02.488884 13912 addons.go:431] installing /etc/kubernetes/addons/helm-tiller-svc.yaml
I0906 18:30:02.488970 13912 ssh_runner.go:362] scp helm-tiller/helm-tiller-svc.yaml --> /etc/kubernetes/addons/helm-tiller-svc.yaml (951 bytes)
I0906 18:30:02.497450 13912 addons.go:431] installing /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml
I0906 18:30:02.497527 13912 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-health-monitor-controller.yaml --> /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml (3038 bytes)
I0906 18:30:02.585190 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml
I0906 18:30:02.688365 13912 addons.go:431] installing /etc/kubernetes/addons/yakd-svc.yaml
I0906 18:30:02.688451 13912 ssh_runner.go:362] scp yakd/yakd-svc.yaml --> /etc/kubernetes/addons/yakd-svc.yaml (412 bytes)
I0906 18:30:02.694691 13912 addons.go:431] installing /etc/kubernetes/addons/registry-proxy.yaml
I0906 18:30:02.694786 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/registry-proxy.yaml (947 bytes)
I0906 18:30:02.776466 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/deployment.yaml
I0906 18:30:02.779398 13912 addons.go:431] installing /etc/kubernetes/addons/ig-role.yaml
I0906 18:30:02.779466 13912 ssh_runner.go:362] scp inspektor-gadget/ig-role.yaml --> /etc/kubernetes/addons/ig-role.yaml (210 bytes)
I0906 18:30:02.879639 13912 addons.go:431] installing /etc/kubernetes/addons/rbac-external-provisioner.yaml
I0906 18:30:02.879733 13912 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-provisioner.yaml --> /etc/kubernetes/addons/rbac-external-provisioner.yaml (4442 bytes)
I0906 18:30:02.973928 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml
I0906 18:30:02.976492 13912 addons.go:431] installing /etc/kubernetes/addons/metrics-server-rbac.yaml
I0906 18:30:02.976572 13912 ssh_runner.go:362] scp metrics-server/metrics-server-rbac.yaml --> /etc/kubernetes/addons/metrics-server-rbac.yaml (2175 bytes)
I0906 18:30:02.978282 13912 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml
I0906 18:30:02.978352 13912 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshotcontents.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml (23126 bytes)
I0906 18:30:03.386971 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml
I0906 18:30:03.388415 13912 addons.go:431] installing /etc/kubernetes/addons/metrics-server-service.yaml
I0906 18:30:03.388472 13912 ssh_runner.go:362] scp metrics-server/metrics-server-service.yaml --> /etc/kubernetes/addons/metrics-server-service.yaml (446 bytes)
I0906 18:30:03.474847 13912 addons.go:431] installing /etc/kubernetes/addons/yakd-dp.yaml
I0906 18:30:03.474938 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/yakd-dp.yaml (2017 bytes)
I0906 18:30:03.480625 13912 addons.go:431] installing /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml
I0906 18:30:03.480704 13912 ssh_runner.go:362] scp volumesnapshots/snapshot.storage.k8s.io_volumesnapshots.yaml --> /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml (19582 bytes)
I0906 18:30:03.481163 13912 addons.go:431] installing /etc/kubernetes/addons/ig-rolebinding.yaml
I0906 18:30:03.481234 13912 ssh_runner.go:362] scp inspektor-gadget/ig-rolebinding.yaml --> /etc/kubernetes/addons/ig-rolebinding.yaml (244 bytes)
I0906 18:30:03.674584 13912 addons.go:431] installing /etc/kubernetes/addons/rbac-external-resizer.yaml
I0906 18:30:03.674613 13912 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-resizer.yaml --> /etc/kubernetes/addons/rbac-external-resizer.yaml (2943 bytes)
I0906 18:30:03.773495 13912 addons.go:431] installing /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml
I0906 18:30:03.773584 13912 ssh_runner.go:362] scp volumesnapshots/rbac-volume-snapshot-controller.yaml --> /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml (3545 bytes)
I0906 18:30:03.892245 13912 ssh_runner.go:235] Completed: sudo systemctl start kubelet: (1.994235856s)
I0906 18:30:03.892342 13912 ssh_runner.go:235] Completed: /bin/bash -c "sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig -n kube-system get configmap coredns -o yaml | sed -e '/^ forward . \/etc\/resolv.conf.*/i \ hosts {\n 192.168.49.1 host.minikube.internal\n fallthrough\n }' -e '/^ errors *$/i \ log' | sudo /var/lib/minikube/binaries/v1.31.0/kubectl --kubeconfig=/var/lib/minikube/kubeconfig replace -f -": (1.994572872s)
I0906 18:30:03.892516 13912 start.go:971] {"host.minikube.internal": 192.168.49.1} host record injected into CoreDNS's ConfigMap
I0906 18:30:03.894498 13912 node_ready.go:35] waiting up to 6m0s for node "addons-141392" to be "Ready" ...
I0906 18:30:03.975488 13912 node_ready.go:49] node "addons-141392" has status "Ready":"True"
I0906 18:30:03.975552 13912 node_ready.go:38] duration metric: took 80.906379ms for node "addons-141392" to be "Ready" ...
I0906 18:30:03.975567 13912 pod_ready.go:36] extra waiting up to 6m0s for all system-critical pods including labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0906 18:30:03.984967 13912 pod_ready.go:79] waiting up to 6m0s for pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace to be "Ready" ...
I0906 18:30:04.073515 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml
I0906 18:30:04.172697 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml
I0906 18:30:04.276420 13912 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrole.yaml
I0906 18:30:04.276499 13912 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrole.yaml --> /etc/kubernetes/addons/ig-clusterrole.yaml (1485 bytes)
I0906 18:30:04.476162 13912 kapi.go:214] "coredns" deployment in "kube-system" namespace and "addons-141392" context rescaled to 1 replicas
I0906 18:30:04.676966 13912 addons.go:431] installing /etc/kubernetes/addons/ig-clusterrolebinding.yaml
I0906 18:30:04.677042 13912 ssh_runner.go:362] scp inspektor-gadget/ig-clusterrolebinding.yaml --> /etc/kubernetes/addons/ig-clusterrolebinding.yaml (274 bytes)
I0906 18:30:04.679590 13912 addons.go:431] installing /etc/kubernetes/addons/rbac-external-snapshotter.yaml
I0906 18:30:04.679614 13912 ssh_runner.go:362] scp csi-hostpath-driver/rbac/rbac-external-snapshotter.yaml --> /etc/kubernetes/addons/rbac-external-snapshotter.yaml (3149 bytes)
I0906 18:30:04.681470 13912 addons.go:431] installing /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0906 18:30:04.681489 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml (1475 bytes)
I0906 18:30:05.093229 13912 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-attacher.yaml
I0906 18:30:05.093256 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-attacher.yaml (2143 bytes)
I0906 18:30:05.172907 13912 addons.go:431] installing /etc/kubernetes/addons/ig-crd.yaml
I0906 18:30:05.172988 13912 ssh_runner.go:362] scp inspektor-gadget/ig-crd.yaml --> /etc/kubernetes/addons/ig-crd.yaml (5216 bytes)
I0906 18:30:05.173510 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0906 18:30:05.876688 13912 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml
I0906 18:30:05.876718 13912 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-driverinfo.yaml --> /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml (1274 bytes)
I0906 18:30:05.992038 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:06.382052 13912 addons.go:431] installing /etc/kubernetes/addons/ig-daemonset.yaml
I0906 18:30:06.382087 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/ig-daemonset.yaml (7735 bytes)
I0906 18:30:06.486604 13912 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-plugin.yaml
I0906 18:30:06.486634 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-plugin.yaml (8201 bytes)
I0906 18:30:06.589284 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner.yaml: (4.512184592s)
I0906 18:30:06.589386 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-dns-pod.yaml: (4.503065698s)
I0906 18:30:06.589426 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/nvidia-device-plugin.yaml: (4.396962741s)
I0906 18:30:07.087319 13912 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-resizer.yaml
I0906 18:30:07.087407 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/csi-hostpath-resizer.yaml (2191 bytes)
I0906 18:30:07.289965 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml
I0906 18:30:07.394055 13912 addons.go:431] installing /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0906 18:30:07.394133 13912 ssh_runner.go:362] scp csi-hostpath-driver/deploy/csi-hostpath-storageclass.yaml --> /etc/kubernetes/addons/csi-hostpath-storageclass.yaml (846 bytes)
I0906 18:30:07.692233 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml
I0906 18:30:07.992118 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:08.578742 13912 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_application_credentials.json (162 bytes)
I0906 18:30:08.578828 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:08.606265 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:09.690836 13912 ssh_runner.go:362] scp memory --> /var/lib/minikube/google_cloud_project (12 bytes)
I0906 18:30:09.994988 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:10.086982 13912 addons.go:234] Setting addon gcp-auth=true in "addons-141392"
I0906 18:30:10.087055 13912 host.go:66] Checking if "addons-141392" exists ...
I0906 18:30:10.087609 13912 cli_runner.go:164] Run: docker container inspect addons-141392 --format={{.State.Status}}
I0906 18:30:10.108131 13912 ssh_runner.go:195] Run: cat /var/lib/minikube/google_application_credentials.json
I0906 18:30:10.108190 13912 cli_runner.go:164] Run: docker container inspect -f "'{{(index (index .NetworkSettings.Ports "22/tcp") 0).HostPort}}'" addons-141392
I0906 18:30:10.123722 13912 sshutil.go:53] new ssh client: &{IP:127.0.0.1 Port:32768 SSHKeyPath:/home/jenkins/minikube-integration/19576-5848/.minikube/machines/addons-141392/id_rsa Username:docker}
I0906 18:30:11.995959 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:13.978100 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/volcano-deployment.yaml: (11.784726647s)
I0906 18:30:13.978316 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ingress-deploy.yaml: (11.705541888s)
I0906 18:30:13.978364 13912 addons.go:475] Verifying addon ingress=true in "addons-141392"
I0906 18:30:13.978431 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storageclass.yaml: (11.68619911s)
I0906 18:30:13.978531 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/storage-provisioner-rancher.yaml: (11.393244239s)
I0906 18:30:13.978669 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/deployment.yaml: (11.202108271s)
I0906 18:30:13.978727 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/helm-tiller-dp.yaml -f /etc/kubernetes/addons/helm-tiller-rbac.yaml -f /etc/kubernetes/addons/helm-tiller-svc.yaml: (11.004719777s)
I0906 18:30:13.978764 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/registry-rc.yaml -f /etc/kubernetes/addons/registry-svc.yaml -f /etc/kubernetes/addons/registry-proxy.yaml: (10.591690559s)
I0906 18:30:13.978791 13912 addons.go:475] Verifying addon registry=true in "addons-141392"
I0906 18:30:13.978966 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/metrics-apiservice.yaml -f /etc/kubernetes/addons/metrics-server-deployment.yaml -f /etc/kubernetes/addons/metrics-server-rbac.yaml -f /etc/kubernetes/addons/metrics-server-service.yaml: (9.905398774s)
I0906 18:30:13.978982 13912 addons.go:475] Verifying addon metrics-server=true in "addons-141392"
I0906 18:30:13.979037 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/yakd-ns.yaml -f /etc/kubernetes/addons/yakd-sa.yaml -f /etc/kubernetes/addons/yakd-crb.yaml -f /etc/kubernetes/addons/yakd-svc.yaml -f /etc/kubernetes/addons/yakd-dp.yaml: (9.806253664s)
I0906 18:30:13.979441 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/ig-namespace.yaml -f /etc/kubernetes/addons/ig-serviceaccount.yaml -f /etc/kubernetes/addons/ig-role.yaml -f /etc/kubernetes/addons/ig-rolebinding.yaml -f /etc/kubernetes/addons/ig-clusterrole.yaml -f /etc/kubernetes/addons/ig-clusterrolebinding.yaml -f /etc/kubernetes/addons/ig-crd.yaml -f /etc/kubernetes/addons/ig-daemonset.yaml: (6.689367571s)
I0906 18:30:13.979305 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (8.805725412s)
W0906 18:30:13.979913 13912 addons.go:457] apply failed, will retry: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0906 18:30:13.979939 13912 retry.go:31] will retry after 276.461012ms: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: Process exited with status 1
stdout:
customresourcedefinition.apiextensions.k8s.io/volumesnapshotclasses.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshotcontents.snapshot.storage.k8s.io created
customresourcedefinition.apiextensions.k8s.io/volumesnapshots.snapshot.storage.k8s.io created
serviceaccount/snapshot-controller created
clusterrole.rbac.authorization.k8s.io/snapshot-controller-runner created
clusterrolebinding.rbac.authorization.k8s.io/snapshot-controller-role created
role.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
rolebinding.rbac.authorization.k8s.io/snapshot-controller-leaderelection created
deployment.apps/snapshot-controller created
stderr:
error: resource mapping not found for name: "csi-hostpath-snapclass" namespace: "" from "/etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml": no matches for kind "VolumeSnapshotClass" in version "snapshot.storage.k8s.io/v1"
ensure CRDs are installed first
I0906 18:30:13.980138 13912 out.go:177] * Verifying registry addon...
I0906 18:30:13.980158 13912 out.go:177] * Verifying ingress addon...
I0906 18:30:13.981086 13912 out.go:177] * To access YAKD - Kubernetes Dashboard, wait for Pod to be ready and run the following command:
minikube -p addons-141392 service yakd-dashboard -n yakd-dashboard
I0906 18:30:13.983136 13912 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=registry" in ns "kube-system" ...
I0906 18:30:13.983859 13912 kapi.go:75] Waiting for pod with label "app.kubernetes.io/name=ingress-nginx" in ns "ingress-nginx" ...
W0906 18:30:13.991049 13912 out.go:270] ! Enabling 'default-storageclass' returned an error: running callbacks: [Error making standard the default storage class: Error while marking storage class local-path as non-default: Operation cannot be fulfilled on storageclasses.storage.k8s.io "local-path": the object has been modified; please apply your changes to the latest version and try again]
I0906 18:30:13.992186 13912 kapi.go:86] Found 2 Pods for label selector kubernetes.io/minikube-addons=registry
I0906 18:30:13.992210 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:13.992669 13912 kapi.go:86] Found 3 Pods for label selector app.kubernetes.io/name=ingress-nginx
I0906 18:30:13.992688 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:13.997294 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:14.257043 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml
I0906 18:30:14.491795 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:14.492221 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:14.987689 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:15.087946 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:15.490505 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:15.491094 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:15.500485 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/rbac-external-attacher.yaml -f /etc/kubernetes/addons/rbac-hostpath.yaml -f /etc/kubernetes/addons/rbac-external-health-monitor-controller.yaml -f /etc/kubernetes/addons/rbac-external-provisioner.yaml -f /etc/kubernetes/addons/rbac-external-resizer.yaml -f /etc/kubernetes/addons/rbac-external-snapshotter.yaml -f /etc/kubernetes/addons/csi-hostpath-attacher.yaml -f /etc/kubernetes/addons/csi-hostpath-driverinfo.yaml -f /etc/kubernetes/addons/csi-hostpath-plugin.yaml -f /etc/kubernetes/addons/csi-hostpath-resizer.yaml -f /etc/kubernetes/addons/csi-hostpath-storageclass.yaml: (7.808190402s)
I0906 18:30:15.500527 13912 addons.go:475] Verifying addon csi-hostpath-driver=true in "addons-141392"
I0906 18:30:15.500698 13912 ssh_runner.go:235] Completed: cat /var/lib/minikube/google_application_credentials.json: (5.392540975s)
I0906 18:30:15.502077 13912 out.go:177] * Verifying csi-hostpath-driver addon...
I0906 18:30:15.502815 13912 out.go:177] - Using image registry.k8s.io/ingress-nginx/kube-webhook-certgen:v1.4.3
I0906 18:30:15.504411 13912 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=csi-hostpath-driver" in ns "kube-system" ...
I0906 18:30:15.505829 13912 out.go:177] - Using image gcr.io/k8s-minikube/gcp-auth-webhook:v0.1.2
I0906 18:30:15.506787 13912 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-ns.yaml
I0906 18:30:15.506818 13912 ssh_runner.go:362] scp gcp-auth/gcp-auth-ns.yaml --> /etc/kubernetes/addons/gcp-auth-ns.yaml (700 bytes)
I0906 18:30:15.576475 13912 kapi.go:86] Found 3 Pods for label selector kubernetes.io/minikube-addons=csi-hostpath-driver
I0906 18:30:15.576515 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:15.593467 13912 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-service.yaml
I0906 18:30:15.593493 13912 ssh_runner.go:362] scp gcp-auth/gcp-auth-service.yaml --> /etc/kubernetes/addons/gcp-auth-service.yaml (788 bytes)
I0906 18:30:15.679369 13912 addons.go:431] installing /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0906 18:30:15.679392 13912 ssh_runner.go:362] scp memory --> /etc/kubernetes/addons/gcp-auth-webhook.yaml (5421 bytes)
I0906 18:30:15.702889 13912 ssh_runner.go:195] Run: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml
I0906 18:30:15.987676 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:15.988735 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:16.075975 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:16.487725 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:16.488719 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:16.490732 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:16.575990 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:16.580461 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply --force -f /etc/kubernetes/addons/csi-hostpath-snapshotclass.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotclasses.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshotcontents.yaml -f /etc/kubernetes/addons/snapshot.storage.k8s.io_volumesnapshots.yaml -f /etc/kubernetes/addons/rbac-volume-snapshot-controller.yaml -f /etc/kubernetes/addons/volume-snapshot-controller-deployment.yaml: (2.323371504s)
I0906 18:30:16.986880 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:16.988044 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:17.090243 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:17.184336 13912 ssh_runner.go:235] Completed: sudo KUBECONFIG=/var/lib/minikube/kubeconfig /var/lib/minikube/binaries/v1.31.0/kubectl apply -f /etc/kubernetes/addons/gcp-auth-ns.yaml -f /etc/kubernetes/addons/gcp-auth-service.yaml -f /etc/kubernetes/addons/gcp-auth-webhook.yaml: (1.481361008s)
I0906 18:30:17.186144 13912 addons.go:475] Verifying addon gcp-auth=true in "addons-141392"
I0906 18:30:17.187522 13912 out.go:177] * Verifying gcp-auth addon...
I0906 18:30:17.189416 13912 kapi.go:75] Waiting for pod with label "kubernetes.io/minikube-addons=gcp-auth" in ns "gcp-auth" ...
I0906 18:30:17.191575 13912 kapi.go:86] Found 0 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0906 18:30:17.487086 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:17.487709 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:17.509245 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:17.986893 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:17.987513 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:18.008820 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:18.486457 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:18.487348 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:18.508801 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:18.986898 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:18.988051 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:18.989963 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:19.008810 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:19.487592 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:19.487885 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:19.508967 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:19.987210 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:19.987684 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:20.007471 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:20.486231 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:20.487353 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:20.508603 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:20.987538 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:20.987745 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:21.007991 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:21.486891 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:21.487203 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:21.490099 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:21.508897 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:21.987356 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:21.987769 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:22.007814 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:22.487796 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:22.488174 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:22.508988 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:22.987376 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:22.988217 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:23.007979 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:23.486789 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:23.488009 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:23.491129 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:23.508385 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:23.987353 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:23.987631 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:24.008577 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:24.486924 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:24.487251 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:24.507932 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:24.987229 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:24.987892 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:25.008576 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:25.487175 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:25.487784 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:25.508037 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:25.986990 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:25.987456 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:25.990222 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:26.008829 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:26.486956 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:26.487430 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:26.507932 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:26.987370 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:26.987787 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:27.009482 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:27.486840 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:27.487993 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:27.508356 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:27.987611 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:27.988588 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:27.990567 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:28.008799 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:28.487107 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:28.487599 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:28.588468 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:28.987825 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:28.988361 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:29.008174 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:29.487274 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:29.488219 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:29.508623 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:29.987056 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:29.987715 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:30.008107 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:30.487353 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:30.487777 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:30.489183 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:30.508463 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:30.987232 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:30.988404 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:31.008391 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:31.487365 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:31.487933 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:31.507823 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:31.986875 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:31.988009 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:32.007920 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:32.488395 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:32.488657 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:32.490997 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:32.508514 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:32.988150 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:32.988606 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:33.008965 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:33.486493 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:33.487751 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:33.508607 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:33.986476 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:33.987435 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:34.011343 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:34.486556 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:34.487375 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:34.509120 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:34.986839 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:34.987651 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:34.989783 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:35.007889 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:35.487363 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=registry", current state: Pending: [<nil>]
I0906 18:30:35.487835 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:35.509084 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:35.988718 13912 kapi.go:107] duration metric: took 22.005578105s to wait for kubernetes.io/minikube-addons=registry ...
I0906 18:30:35.990469 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:36.092186 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:36.487827 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:36.508941 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:36.988872 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:36.990408 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:37.090186 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:37.488375 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:37.508549 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:37.987816 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:38.008955 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:38.487820 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:38.509161 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:38.988725 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:39.008751 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:39.488708 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:39.491583 13912 pod_ready.go:103] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"False"
I0906 18:30:39.509695 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:39.987935 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:39.989682 13912 pod_ready.go:93] pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace has status "Ready":"True"
I0906 18:30:39.989699 13912 pod_ready.go:82] duration metric: took 36.004653654s for pod "coredns-6f6b679f8f-kqzqs" in "kube-system" namespace to be "Ready" ...
I0906 18:30:39.989708 13912 pod_ready.go:79] waiting up to 6m0s for pod "coredns-6f6b679f8f-tgs62" in "kube-system" namespace to be "Ready" ...
I0906 18:30:39.991164 13912 pod_ready.go:98] error getting pod "coredns-6f6b679f8f-tgs62" in "kube-system" namespace (skipping!): pods "coredns-6f6b679f8f-tgs62" not found
I0906 18:30:39.991187 13912 pod_ready.go:82] duration metric: took 1.471542ms for pod "coredns-6f6b679f8f-tgs62" in "kube-system" namespace to be "Ready" ...
E0906 18:30:39.991198 13912 pod_ready.go:67] WaitExtra: waitPodCondition: error getting pod "coredns-6f6b679f8f-tgs62" in "kube-system" namespace (skipping!): pods "coredns-6f6b679f8f-tgs62" not found
I0906 18:30:39.991207 13912 pod_ready.go:79] waiting up to 6m0s for pod "etcd-addons-141392" in "kube-system" namespace to be "Ready" ...
I0906 18:30:39.995280 13912 pod_ready.go:93] pod "etcd-addons-141392" in "kube-system" namespace has status "Ready":"True"
I0906 18:30:39.995299 13912 pod_ready.go:82] duration metric: took 4.080754ms for pod "etcd-addons-141392" in "kube-system" namespace to be "Ready" ...
I0906 18:30:39.995310 13912 pod_ready.go:79] waiting up to 6m0s for pod "kube-apiserver-addons-141392" in "kube-system" namespace to be "Ready" ...
I0906 18:30:39.999365 13912 pod_ready.go:93] pod "kube-apiserver-addons-141392" in "kube-system" namespace has status "Ready":"True"
I0906 18:30:39.999389 13912 pod_ready.go:82] duration metric: took 4.06973ms for pod "kube-apiserver-addons-141392" in "kube-system" namespace to be "Ready" ...
I0906 18:30:39.999403 13912 pod_ready.go:79] waiting up to 6m0s for pod "kube-controller-manager-addons-141392" in "kube-system" namespace to be "Ready" ...
I0906 18:30:40.003103 13912 pod_ready.go:93] pod "kube-controller-manager-addons-141392" in "kube-system" namespace has status "Ready":"True"
I0906 18:30:40.003123 13912 pod_ready.go:82] duration metric: took 3.711101ms for pod "kube-controller-manager-addons-141392" in "kube-system" namespace to be "Ready" ...
I0906 18:30:40.003134 13912 pod_ready.go:79] waiting up to 6m0s for pod "kube-proxy-dhft7" in "kube-system" namespace to be "Ready" ...
I0906 18:30:40.007324 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:40.188564 13912 pod_ready.go:93] pod "kube-proxy-dhft7" in "kube-system" namespace has status "Ready":"True"
I0906 18:30:40.188589 13912 pod_ready.go:82] duration metric: took 185.446803ms for pod "kube-proxy-dhft7" in "kube-system" namespace to be "Ready" ...
I0906 18:30:40.188601 13912 pod_ready.go:79] waiting up to 6m0s for pod "kube-scheduler-addons-141392" in "kube-system" namespace to be "Ready" ...
I0906 18:30:40.487856 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:40.509118 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:40.589029 13912 pod_ready.go:93] pod "kube-scheduler-addons-141392" in "kube-system" namespace has status "Ready":"True"
I0906 18:30:40.589057 13912 pod_ready.go:82] duration metric: took 400.446171ms for pod "kube-scheduler-addons-141392" in "kube-system" namespace to be "Ready" ...
I0906 18:30:40.589073 13912 pod_ready.go:79] waiting up to 6m0s for pod "nvidia-device-plugin-daemonset-rhlhv" in "kube-system" namespace to be "Ready" ...
I0906 18:30:40.987809 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:40.988459 13912 pod_ready.go:93] pod "nvidia-device-plugin-daemonset-rhlhv" in "kube-system" namespace has status "Ready":"True"
I0906 18:30:40.988483 13912 pod_ready.go:82] duration metric: took 399.40049ms for pod "nvidia-device-plugin-daemonset-rhlhv" in "kube-system" namespace to be "Ready" ...
I0906 18:30:40.988493 13912 pod_ready.go:39] duration metric: took 37.012910055s for extra waiting for all system-critical and pods with labels [k8s-app=kube-dns component=etcd component=kube-apiserver component=kube-controller-manager k8s-app=kube-proxy component=kube-scheduler] to be "Ready" ...
I0906 18:30:40.988517 13912 api_server.go:52] waiting for apiserver process to appear ...
I0906 18:30:40.988583 13912 ssh_runner.go:195] Run: sudo pgrep -xnf kube-apiserver.*minikube.*
I0906 18:30:41.005836 13912 api_server.go:72] duration metric: took 39.517782007s to wait for apiserver process to appear ...
I0906 18:30:41.005867 13912 api_server.go:88] waiting for apiserver healthz status ...
I0906 18:30:41.005886 13912 api_server.go:253] Checking apiserver healthz at https://192.168.49.2:8443/healthz ...
I0906 18:30:41.008713 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:41.009909 13912 api_server.go:279] https://192.168.49.2:8443/healthz returned 200:
ok
I0906 18:30:41.010669 13912 api_server.go:141] control plane version: v1.31.0
I0906 18:30:41.010691 13912 api_server.go:131] duration metric: took 4.815871ms to wait for apiserver health ...
I0906 18:30:41.010704 13912 system_pods.go:43] waiting for kube-system pods to appear ...
I0906 18:30:41.195381 13912 system_pods.go:59] 18 kube-system pods found
I0906 18:30:41.195409 13912 system_pods.go:61] "coredns-6f6b679f8f-kqzqs" [e005ec77-8350-41fa-adcd-2714fd9301a6] Running
I0906 18:30:41.195417 13912 system_pods.go:61] "csi-hostpath-attacher-0" [b566aed3-a147-4f79-a723-32f6986289af] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0906 18:30:41.195423 13912 system_pods.go:61] "csi-hostpath-resizer-0" [06cd6dcf-3dd2-4bf0-8c03-5d7014fa42d4] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0906 18:30:41.195431 13912 system_pods.go:61] "csi-hostpathplugin-njtzf" [b281453b-0c25-4436-bbd8-b21cbbafd846] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0906 18:30:41.195439 13912 system_pods.go:61] "etcd-addons-141392" [c37a5390-5f25-4f57-a920-8d0870ac52ea] Running
I0906 18:30:41.195445 13912 system_pods.go:61] "kube-apiserver-addons-141392" [14d43d92-5197-4161-9ba6-4408c3b3bc3d] Running
I0906 18:30:41.195454 13912 system_pods.go:61] "kube-controller-manager-addons-141392" [98f1b865-aa5e-44df-bf3e-98ab561f7438] Running
I0906 18:30:41.195466 13912 system_pods.go:61] "kube-ingress-dns-minikube" [167a4c1d-ef6a-46bd-8549-54be669f20cb] Running
I0906 18:30:41.195471 13912 system_pods.go:61] "kube-proxy-dhft7" [acf3c0c9-31dc-4b8c-bec8-febd27e21190] Running
I0906 18:30:41.195479 13912 system_pods.go:61] "kube-scheduler-addons-141392" [fd4d2d87-b781-4970-8ff8-75abab51bbb0] Running
I0906 18:30:41.195487 13912 system_pods.go:61] "metrics-server-84c5f94fbc-6xvn9" [9f36f9dc-f790-4477-844d-c402d851062c] Running / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0906 18:30:41.195497 13912 system_pods.go:61] "nvidia-device-plugin-daemonset-rhlhv" [5c691bb8-c1de-48ba-b9ec-b5e084831471] Running
I0906 18:30:41.195506 13912 system_pods.go:61] "registry-6fb4cdfc84-d989j" [03e8d3f6-b271-4c1d-a933-77656192d8eb] Running
I0906 18:30:41.195511 13912 system_pods.go:61] "registry-proxy-c7665" [23346424-7dc9-4199-af57-8fffc4bcfe51] Running
I0906 18:30:41.195531 13912 system_pods.go:61] "snapshot-controller-56fcc65765-6w2zf" [77c00cce-b25f-434d-80ef-7ac48ccca8e0] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0906 18:30:41.195539 13912 system_pods.go:61] "snapshot-controller-56fcc65765-bl8fr" [38ddbac3-cf56-4d53-b90e-413e63a646e5] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0906 18:30:41.195545 13912 system_pods.go:61] "storage-provisioner" [1b01a828-baf6-48ba-9771-271ccf44ded4] Running
I0906 18:30:41.195557 13912 system_pods.go:61] "tiller-deploy-b48cc5f79-mj9zj" [68e64cb3-d2a3-4d2a-9093-ef2fb97a0fbb] Running
I0906 18:30:41.195567 13912 system_pods.go:74] duration metric: took 184.85693ms to wait for pod list to return data ...
I0906 18:30:41.195579 13912 default_sa.go:34] waiting for default service account to be created ...
I0906 18:30:41.388334 13912 default_sa.go:45] found service account: "default"
I0906 18:30:41.388362 13912 default_sa.go:55] duration metric: took 192.775785ms for default service account to be created ...
I0906 18:30:41.388373 13912 system_pods.go:116] waiting for k8s-apps to be running ...
I0906 18:30:41.488430 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:41.509893 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:41.595381 13912 system_pods.go:86] 18 kube-system pods found
I0906 18:30:41.595416 13912 system_pods.go:89] "coredns-6f6b679f8f-kqzqs" [e005ec77-8350-41fa-adcd-2714fd9301a6] Running
I0906 18:30:41.595431 13912 system_pods.go:89] "csi-hostpath-attacher-0" [b566aed3-a147-4f79-a723-32f6986289af] Pending / Ready:ContainersNotReady (containers with unready status: [csi-attacher]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-attacher])
I0906 18:30:41.595439 13912 system_pods.go:89] "csi-hostpath-resizer-0" [06cd6dcf-3dd2-4bf0-8c03-5d7014fa42d4] Pending / Ready:ContainersNotReady (containers with unready status: [csi-resizer]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-resizer])
I0906 18:30:41.595450 13912 system_pods.go:89] "csi-hostpathplugin-njtzf" [b281453b-0c25-4436-bbd8-b21cbbafd846] Pending / Ready:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter]) / ContainersReady:ContainersNotReady (containers with unready status: [csi-external-health-monitor-controller node-driver-registrar hostpath liveness-probe csi-provisioner csi-snapshotter])
I0906 18:30:41.595456 13912 system_pods.go:89] "etcd-addons-141392" [c37a5390-5f25-4f57-a920-8d0870ac52ea] Running
I0906 18:30:41.595466 13912 system_pods.go:89] "kube-apiserver-addons-141392" [14d43d92-5197-4161-9ba6-4408c3b3bc3d] Running
I0906 18:30:41.595472 13912 system_pods.go:89] "kube-controller-manager-addons-141392" [98f1b865-aa5e-44df-bf3e-98ab561f7438] Running
I0906 18:30:41.595481 13912 system_pods.go:89] "kube-ingress-dns-minikube" [167a4c1d-ef6a-46bd-8549-54be669f20cb] Running
I0906 18:30:41.595486 13912 system_pods.go:89] "kube-proxy-dhft7" [acf3c0c9-31dc-4b8c-bec8-febd27e21190] Running
I0906 18:30:41.595491 13912 system_pods.go:89] "kube-scheduler-addons-141392" [fd4d2d87-b781-4970-8ff8-75abab51bbb0] Running
I0906 18:30:41.595502 13912 system_pods.go:89] "metrics-server-84c5f94fbc-6xvn9" [9f36f9dc-f790-4477-844d-c402d851062c] Running / Ready:ContainersNotReady (containers with unready status: [metrics-server]) / ContainersReady:ContainersNotReady (containers with unready status: [metrics-server])
I0906 18:30:41.595512 13912 system_pods.go:89] "nvidia-device-plugin-daemonset-rhlhv" [5c691bb8-c1de-48ba-b9ec-b5e084831471] Running
I0906 18:30:41.595549 13912 system_pods.go:89] "registry-6fb4cdfc84-d989j" [03e8d3f6-b271-4c1d-a933-77656192d8eb] Running
I0906 18:30:41.595555 13912 system_pods.go:89] "registry-proxy-c7665" [23346424-7dc9-4199-af57-8fffc4bcfe51] Running
I0906 18:30:41.595566 13912 system_pods.go:89] "snapshot-controller-56fcc65765-6w2zf" [77c00cce-b25f-434d-80ef-7ac48ccca8e0] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0906 18:30:41.595578 13912 system_pods.go:89] "snapshot-controller-56fcc65765-bl8fr" [38ddbac3-cf56-4d53-b90e-413e63a646e5] Pending / Ready:ContainersNotReady (containers with unready status: [volume-snapshot-controller]) / ContainersReady:ContainersNotReady (containers with unready status: [volume-snapshot-controller])
I0906 18:30:41.595586 13912 system_pods.go:89] "storage-provisioner" [1b01a828-baf6-48ba-9771-271ccf44ded4] Running
I0906 18:30:41.595592 13912 system_pods.go:89] "tiller-deploy-b48cc5f79-mj9zj" [68e64cb3-d2a3-4d2a-9093-ef2fb97a0fbb] Running
I0906 18:30:41.595602 13912 system_pods.go:126] duration metric: took 207.222495ms to wait for k8s-apps to be running ...
I0906 18:30:41.595611 13912 system_svc.go:44] waiting for kubelet service to be running ....
I0906 18:30:41.595670 13912 ssh_runner.go:195] Run: sudo systemctl is-active --quiet service kubelet
I0906 18:30:41.606574 13912 system_svc.go:56] duration metric: took 10.958114ms WaitForService to wait for kubelet
I0906 18:30:41.606600 13912 kubeadm.go:582] duration metric: took 40.118549182s to wait for: map[apiserver:true apps_running:true default_sa:true extra:true kubelet:true node_ready:true system_pods:true]
I0906 18:30:41.606625 13912 node_conditions.go:102] verifying NodePressure condition ...
I0906 18:30:41.789291 13912 node_conditions.go:122] node storage ephemeral capacity is 304681132Ki
I0906 18:30:41.789318 13912 node_conditions.go:123] node cpu capacity is 8
I0906 18:30:41.789333 13912 node_conditions.go:105] duration metric: took 182.701898ms to run NodePressure ...
I0906 18:30:41.789348 13912 start.go:241] waiting for startup goroutines ...
I0906 18:30:41.987967 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:42.099388 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:42.488475 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:42.508589 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:42.988008 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:43.009130 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:43.493614 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:43.508111 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:43.988196 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:44.008445 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:44.488757 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:44.508802 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:44.987985 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:45.008929 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:45.488573 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:45.508235 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:45.988151 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:46.010091 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:46.488200 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:46.508987 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:46.988416 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:47.089628 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:47.487290 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:47.508693 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:47.988896 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:48.008829 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:48.531917 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:48.547817 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:48.987322 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:49.008528 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:49.489236 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:49.509592 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:49.987292 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:50.008550 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:50.488090 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:50.509305 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:50.987424 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:51.008569 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:51.488565 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:51.508613 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:52.004720 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:52.007792 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:52.488628 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:52.509978 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:52.988247 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:53.008093 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:53.488310 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:53.509615 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:53.988109 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:54.009137 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:54.488519 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:54.507901 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:54.988477 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:55.009319 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:55.487708 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:55.508248 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:55.990022 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:56.092439 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:56.488374 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:56.507884 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:56.988449 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:57.008506 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:57.488261 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:57.508739 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:57.988014 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:58.008863 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:58.487631 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:58.507935 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=csi-hostpath-driver", current state: Pending: [<nil>]
I0906 18:30:58.988275 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:59.008509 13912 kapi.go:107] duration metric: took 43.504097731s to wait for kubernetes.io/minikube-addons=csi-hostpath-driver ...
I0906 18:30:59.488320 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:30:59.987490 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:00.487917 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:00.987454 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:01.488004 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:01.986907 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:02.487327 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:02.987013 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:03.487453 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:03.987444 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:04.488251 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:04.987094 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:05.487490 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:05.987800 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:06.487948 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:06.987228 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:07.487293 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:07.987965 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:08.487705 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:08.988268 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:09.487101 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:09.987791 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:10.487791 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:10.988396 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:11.488205 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:11.988016 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:12.488091 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:12.987063 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:13.487938 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:13.987462 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:14.487478 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:14.987751 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:15.488170 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:15.988653 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:16.488787 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:16.987955 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:17.487774 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:17.988513 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:18.488110 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:18.987706 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:19.488436 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:19.988965 13912 kapi.go:96] waiting for pod "app.kubernetes.io/name=ingress-nginx", current state: Pending: [<nil>]
I0906 18:31:20.487994 13912 kapi.go:107] duration metric: took 1m6.504132862s to wait for app.kubernetes.io/name=ingress-nginx ...
I0906 18:31:40.692741 13912 kapi.go:86] Found 1 Pods for label selector kubernetes.io/minikube-addons=gcp-auth
I0906 18:31:40.692767 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:41.192350 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:41.692776 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:42.192533 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:42.692024 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:43.192885 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:43.692462 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:44.192545 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:44.692533 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:45.192264 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:45.692376 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:46.192466 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:46.692239 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:47.193227 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:47.693287 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:48.192193 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:48.692840 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:49.192730 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:49.692827 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:50.192416 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:50.692184 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:51.192444 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:51.692486 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:52.192334 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:52.692269 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:53.192922 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:53.692376 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:54.191976 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:54.692962 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:55.192611 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:55.692696 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:56.192240 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:56.693798 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:57.192615 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:57.692061 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:58.192883 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:58.692421 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:59.192107 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:31:59.692243 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:00.193410 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:00.692415 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:01.192239 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:01.692195 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:02.193321 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:02.692373 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:03.192221 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:03.692959 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:04.192719 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:04.692710 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:05.192262 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:05.693212 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:06.193057 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:06.692977 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:07.193036 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:07.692663 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:08.192333 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:08.692423 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:09.192443 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:09.692529 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:10.192268 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:10.692306 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:11.192916 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:11.693372 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:12.192475 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:12.692995 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:13.192406 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:13.692067 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:14.193174 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:14.693079 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:15.193186 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:15.693193 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:16.193299 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:16.692465 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:17.192013 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:17.692785 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:18.192151 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:18.692735 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:19.192247 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:19.692888 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:20.192436 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:20.692474 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:21.192049 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:21.720608 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:22.192423 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:22.692059 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:23.192872 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:23.692858 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:24.192488 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:24.692171 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:25.193352 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:25.692230 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:26.193078 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:26.692482 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:27.192184 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:27.693470 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:28.192359 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:28.692010 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:29.192386 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:29.693257 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:30.192983 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:30.693083 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:31.193110 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:31.692665 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:32.193038 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:32.692900 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:33.192697 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:33.692413 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:34.192954 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:34.692757 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:35.192710 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:35.692681 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:36.193055 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:36.692942 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:37.192971 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:37.692975 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:38.192887 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:38.692654 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:39.192626 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:39.692637 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:40.192252 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:40.692993 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:41.192892 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:41.693564 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:42.192182 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:42.693366 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:43.192447 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:43.692166 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:44.193276 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:44.692387 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:45.192285 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:45.693373 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:46.192559 13912 kapi.go:96] waiting for pod "kubernetes.io/minikube-addons=gcp-auth", current state: Pending: [<nil>]
I0906 18:32:46.693726 13912 kapi.go:107] duration metric: took 2m29.504307798s to wait for kubernetes.io/minikube-addons=gcp-auth ...
I0906 18:32:46.694966 13912 out.go:177] * Your GCP credentials will now be mounted into every pod created in the addons-141392 cluster.
I0906 18:32:46.696471 13912 out.go:177] * If you don't want your credentials mounted into a specific pod, add a label with the `gcp-auth-skip-secret` key to your pod configuration.
I0906 18:32:46.697562 13912 out.go:177] * If you want existing pods to be mounted with credentials, either recreate them or rerun addons enable with --refresh.
I0906 18:32:46.698754 13912 out.go:177] * Enabled addons: storage-provisioner, ingress-dns, nvidia-device-plugin, volcano, cloud-spanner, helm-tiller, metrics-server, inspektor-gadget, yakd, storage-provisioner-rancher, volumesnapshots, registry, csi-hostpath-driver, ingress, gcp-auth
I0906 18:32:46.699758 13912 addons.go:510] duration metric: took 2m45.211689457s for enable addons: enabled=[storage-provisioner ingress-dns nvidia-device-plugin volcano cloud-spanner helm-tiller metrics-server inspektor-gadget yakd storage-provisioner-rancher volumesnapshots registry csi-hostpath-driver ingress gcp-auth]
I0906 18:32:46.699797 13912 start.go:246] waiting for cluster config update ...
I0906 18:32:46.699816 13912 start.go:255] writing updated cluster config ...
I0906 18:32:46.700044 13912 ssh_runner.go:195] Run: rm -f paused
I0906 18:32:46.820541 13912 start.go:600] kubectl: 1.31.0, cluster: 1.31.0 (minor skew: 0)
I0906 18:32:46.822073 13912 out.go:177] * Done! kubectl is now configured to use "addons-141392" cluster and "default" namespace by default
==> Docker <==
Sep 06 18:42:10 addons-141392 dockerd[1339]: time="2024-09-06T18:42:10.807891284Z" level=info msg="ignoring event" container=6af0ea10c74410af3b1e67e8f36bf033668c6e0ce48e79a55915456d18290502 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:10 addons-141392 dockerd[1339]: time="2024-09-06T18:42:10.904538724Z" level=info msg="ignoring event" container=71d8ebbb21a95e35ae63728cd640c240be9034e42e96e94b725a9e17ad2e5244 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:16 addons-141392 dockerd[1339]: time="2024-09-06T18:42:16.000103510Z" level=info msg="ignoring event" container=c833a70a70b3453a0454aa864106156473b1a705f09b6b03e79166e0c5721238 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:16 addons-141392 dockerd[1339]: time="2024-09-06T18:42:16.122540747Z" level=info msg="ignoring event" container=dc5440aa623b157f91a6903abba7dd0488aea3a777b3ef418a76222d05b887b1 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:16 addons-141392 dockerd[1339]: time="2024-09-06T18:42:16.168319616Z" level=info msg="Attempting next endpoint for pull after error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 06 18:42:16 addons-141392 dockerd[1339]: time="2024-09-06T18:42:16.169711050Z" level=error msg="Handler for POST /v1.43/images/create returned error: Head \"https://gcr.io/v2/k8s-minikube/busybox/manifests/latest\": unauthorized: authentication failed"
Sep 06 18:42:16 addons-141392 dockerd[1339]: time="2024-09-06T18:42:16.785038702Z" level=info msg="ignoring event" container=3bad9cbe25296be27a4a6ba05cf2ed81990bce2da09809db8c2ac15603ec0f84 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:16 addons-141392 dockerd[1339]: time="2024-09-06T18:42:16.786903300Z" level=info msg="ignoring event" container=c0732f341b06fe4914d7a5a9f12e20581a9470445f64445b466ac12d5f204353 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:16 addons-141392 dockerd[1339]: time="2024-09-06T18:42:16.918089801Z" level=info msg="ignoring event" container=d61b92c0f91efbaed44dc03761105ae29ebb697593ed11fdfa513ffee4cf8076 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:16 addons-141392 dockerd[1339]: time="2024-09-06T18:42:16.982976307Z" level=info msg="ignoring event" container=8fa5dd89021557428edf1651f32020e25cc9216d76e5caa994baa0e2e674afda module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:18 addons-141392 cri-dockerd[1604]: time="2024-09-06T18:42:18Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/80781fe72c1a2157fe8ab2270e059c062b2bc2b7dfd8b5c515f9c7b37bfd49eb/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local us-east4-a.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 06 18:42:19 addons-141392 cri-dockerd[1604]: time="2024-09-06T18:42:19Z" level=info msg="Stop pulling image docker.io/nginx:alpine: Status: Downloaded newer image for nginx:alpine"
Sep 06 18:42:23 addons-141392 dockerd[1339]: time="2024-09-06T18:42:23.394672482Z" level=info msg="ignoring event" container=51b5eafad4c6da15b39b3f12bf2aab8f306aebca9ebbeb562cdb2782e6799e63 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:26 addons-141392 cri-dockerd[1604]: time="2024-09-06T18:42:26Z" level=info msg="Will attempt to re-write config file /var/lib/docker/containers/c230d3e14d071e9e930b2dd4f8e129de374111d72c1c49b8e968b3c93cc519d3/resolv.conf as [nameserver 10.96.0.10 search default.svc.cluster.local svc.cluster.local cluster.local us-east4-a.c.k8s-minikube.internal c.k8s-minikube.internal google.internal options ndots:5]"
Sep 06 18:42:26 addons-141392 dockerd[1339]: time="2024-09-06T18:42:26.828577183Z" level=info msg="ignoring event" container=e2e69a87094b2bf605dd6f54af3a00f28faecb92060de546236196e0a99814c7 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:26 addons-141392 dockerd[1339]: time="2024-09-06T18:42:26.888414211Z" level=info msg="ignoring event" container=8c2b8c3000360ca4404b4f8225987788fa25f1d4d69cb146a5ccedac360b60f6 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:26 addons-141392 cri-dockerd[1604]: time="2024-09-06T18:42:26Z" level=info msg="Stop pulling image docker.io/kicbase/echo-server:1.0: Status: Downloaded newer image for kicbase/echo-server:1.0"
Sep 06 18:42:31 addons-141392 dockerd[1339]: time="2024-09-06T18:42:31.598630538Z" level=info msg="Container failed to exit within 2s of signal 15 - using the force" container=5f997566aefdbbbb92ccd8ecefa1508f5c70f76ea26ee3a666e7e6613dd3da04
Sep 06 18:42:31 addons-141392 dockerd[1339]: time="2024-09-06T18:42:31.648642579Z" level=info msg="ignoring event" container=5f997566aefdbbbb92ccd8ecefa1508f5c70f76ea26ee3a666e7e6613dd3da04 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:31 addons-141392 dockerd[1339]: time="2024-09-06T18:42:31.784407749Z" level=info msg="ignoring event" container=952bcbabe5c56ca20008e59874738d9d9901797c63ba6225559543fbc2ee7be2 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:36 addons-141392 dockerd[1339]: time="2024-09-06T18:42:36.895794719Z" level=info msg="ignoring event" container=c7c4fcf43748d2b703fd3124954cf31c8fb22c6bf027f8fe6df7acaa0d56de0d module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:37 addons-141392 dockerd[1339]: time="2024-09-06T18:42:37.388790117Z" level=info msg="ignoring event" container=e05cbbcfa65246fc4b39126a3ba491189b4aefc7b0237ef99f275c231fa08cab module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:37 addons-141392 dockerd[1339]: time="2024-09-06T18:42:37.450373017Z" level=info msg="ignoring event" container=8ea787065b670dd7cdb148ecb0bd08e1cb78613ad289e66cff17497e4b649a32 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:37 addons-141392 dockerd[1339]: time="2024-09-06T18:42:37.524582888Z" level=info msg="ignoring event" container=bcade79f224653c0e39db684dd749edb2d1d9ad81ae88cf10434fa81d93b0369 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
Sep 06 18:42:37 addons-141392 dockerd[1339]: time="2024-09-06T18:42:37.596074019Z" level=info msg="ignoring event" container=65bc7c702f8a94747b7bec67b5314ee68646cf86d3d258058c3cbdeee95f8a01 module=libcontainerd namespace=moby topic=/tasks/delete type="*events.TaskDelete"
==> container status <==
CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID POD
e4d7df39b67d9 kicbase/echo-server@sha256:127ac38a2bb9537b7f252addff209ea6801edcac8a92c8b1104dacd66a583ed6 12 seconds ago Running hello-world-app 0 c230d3e14d071 hello-world-app-55bf9c44b4-lqc8k
50d055f369db5 nginx@sha256:c04c18adc2a407740a397c8407c011fc6c90026a9b65cceddef7ae5484360158 19 seconds ago Running nginx 0 80781fe72c1a2 nginx
e31a51d72f1f4 alpine/helm@sha256:9d9fab00e0680f1328924429925595dfe96a68531c8a9c1518d05ee2ad45c36f 36 seconds ago Exited helm-test 0 d98cdce987723 helm-test
be65c977e4e89 gcr.io/k8s-minikube/gcp-auth-webhook@sha256:e6c5b3bc32072ea370d34c27836efd11b3519d25bd444c2a8efc339cff0e20fb 9 minutes ago Running gcp-auth 0 b995e740fac5e gcp-auth-89d5ffd79-5rxj9
63ff58960910d ce263a8653f9c 11 minutes ago Exited patch 2 e4987c74cd33b ingress-nginx-admission-patch-79mbp
cf7232fe81ad8 registry.k8s.io/ingress-nginx/kube-webhook-certgen@sha256:a320a50cc91bd15fd2d6fa6de58bd98c1bd64b9a6f926ce23a600d87043455a3 11 minutes ago Exited create 0 b403fe7b84853 ingress-nginx-admission-create-qjjs7
f1925f13e8318 6e38f40d628db 12 minutes ago Running storage-provisioner 0 12b22f9e10e05 storage-provisioner
6a0bf988acec5 cbb01a7bd410d 12 minutes ago Running coredns 0 31ccedd6d3366 coredns-6f6b679f8f-kqzqs
7ff2536e76737 ad83b2ca7b09e 12 minutes ago Running kube-proxy 0 6ec25d80d86d8 kube-proxy-dhft7
34baa43f88054 045733566833c 12 minutes ago Running kube-controller-manager 0 5269efa293da2 kube-controller-manager-addons-141392
bd5471ad3706e 1766f54c897f0 12 minutes ago Running kube-scheduler 0 b3f91c12a11e9 kube-scheduler-addons-141392
472ba18d688a8 604f5db92eaa8 12 minutes ago Running kube-apiserver 0 49dfca1e15117 kube-apiserver-addons-141392
41162656997ea 2e96e5913fc06 12 minutes ago Running etcd 0 976515d9ea8c5 etcd-addons-141392
==> coredns [6a0bf988acec] <==
[INFO] 10.244.0.22:43042 - 46953 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.004303281s
[INFO] 10.244.0.22:48553 - 58079 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004799603s
[INFO] 10.244.0.22:40631 - 63595 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004650349s
[INFO] 10.244.0.22:43658 - 22568 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004841113s
[INFO] 10.244.0.22:48096 - 27299 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004849388s
[INFO] 10.244.0.22:48553 - 24124 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.00007s
[INFO] 10.244.0.22:43658 - 11043 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000113121s
[INFO] 10.244.0.22:43042 - 39023 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.062408735s
[INFO] 10.244.0.22:40631 - 5257 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000081516s
[INFO] 10.244.0.22:48096 - 27690 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000141143s
[INFO] 10.244.0.22:54904 - 13569 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.06269724s
[INFO] 10.244.0.22:50077 - 6778 "A IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.006571675s
[INFO] 10.244.0.22:47861 - 18818 "A IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.006785266s
[INFO] 10.244.0.22:43042 - 10640 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.007051666s
[INFO] 10.244.0.22:43042 - 27761 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000068451s
[INFO] 10.244.0.22:54904 - 28044 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004145623s
[INFO] 10.244.0.22:54904 - 34181 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000066728s
[INFO] 10.244.0.22:47861 - 64291 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.003911565s
[INFO] 10.244.0.22:50077 - 33570 "AAAA IN hello-world-app.default.svc.cluster.local.c.k8s-minikube.internal. udp 83 false 512" NXDOMAIN qr,rd,ra 83 0.005425274s
[INFO] 10.244.0.22:47861 - 10194 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004035667s
[INFO] 10.244.0.22:50077 - 21242 "A IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.004058993s
[INFO] 10.244.0.22:47861 - 1169 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005844856s
[INFO] 10.244.0.22:50077 - 11198 "AAAA IN hello-world-app.default.svc.cluster.local.google.internal. udp 75 false 512" NXDOMAIN qr,rd,ra 75 0.005646811s
[INFO] 10.244.0.22:47861 - 52567 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000057789s
[INFO] 10.244.0.22:50077 - 62471 "A IN hello-world-app.default.svc.cluster.local. udp 59 false 512" NOERROR qr,aa,rd 116 0.000066895s
==> describe nodes <==
Name: addons-141392
Roles: control-plane
Labels: beta.kubernetes.io/arch=amd64
beta.kubernetes.io/os=linux
kubernetes.io/arch=amd64
kubernetes.io/hostname=addons-141392
kubernetes.io/os=linux
minikube.k8s.io/commit=e6b6435971a63e36b5096cd544634422129cef13
minikube.k8s.io/name=addons-141392
minikube.k8s.io/primary=true
minikube.k8s.io/updated_at=2024_09_06T18_29_56_0700
minikube.k8s.io/version=v1.34.0
node-role.kubernetes.io/control-plane=
node.kubernetes.io/exclude-from-external-load-balancers=
topology.hostpath.csi/node=addons-141392
Annotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/cri-dockerd.sock
node.alpha.kubernetes.io/ttl: 0
volumes.kubernetes.io/controller-managed-attach-detach: true
CreationTimestamp: Fri, 06 Sep 2024 18:29:54 +0000
Taints: <none>
Unschedulable: false
Lease:
HolderIdentity: addons-141392
AcquireTime: <unset>
RenewTime: Fri, 06 Sep 2024 18:42:30 +0000
Conditions:
Type Status LastHeartbeatTime LastTransitionTime Reason Message
---- ------ ----------------- ------------------ ------ -------
MemoryPressure False Fri, 06 Sep 2024 18:42:30 +0000 Fri, 06 Sep 2024 18:29:52 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available
DiskPressure False Fri, 06 Sep 2024 18:42:30 +0000 Fri, 06 Sep 2024 18:29:52 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure
PIDPressure False Fri, 06 Sep 2024 18:42:30 +0000 Fri, 06 Sep 2024 18:29:52 +0000 KubeletHasSufficientPID kubelet has sufficient PID available
Ready True Fri, 06 Sep 2024 18:42:30 +0000 Fri, 06 Sep 2024 18:29:54 +0000 KubeletReady kubelet is posting ready status
Addresses:
InternalIP: 192.168.49.2
Hostname: addons-141392
Capacity:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
Allocatable:
cpu: 8
ephemeral-storage: 304681132Ki
hugepages-1Gi: 0
hugepages-2Mi: 0
memory: 32859316Ki
pods: 110
System Info:
Machine ID: ab260f75d3004a4683d74c76f34fca4d
System UUID: 93a71428-90b2-4c7b-81b5-1d147e230b23
Boot ID: b085ec8d-bee9-4df0-b598-48b9b817369b
Kernel Version: 5.15.0-1067-gcp
OS Image: Ubuntu 22.04.4 LTS
Operating System: linux
Architecture: amd64
Container Runtime Version: docker://27.2.0
Kubelet Version: v1.31.0
Kube-Proxy Version:
PodCIDR: 10.244.0.0/24
PodCIDRs: 10.244.0.0/24
Non-terminated Pods: (11 in total)
Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age
--------- ---- ------------ ---------- --------------- ------------- ---
default busybox 0 (0%) 0 (0%) 0 (0%) 0 (0%) 9m14s
default hello-world-app-55bf9c44b4-lqc8k 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12s
default nginx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 21s
gcp-auth gcp-auth-89d5ffd79-5rxj9 0 (0%) 0 (0%) 0 (0%) 0 (0%) 10m
kube-system coredns-6f6b679f8f-kqzqs 100m (1%) 0 (0%) 70Mi (0%) 170Mi (0%) 12m
kube-system etcd-addons-141392 100m (1%) 0 (0%) 100Mi (0%) 0 (0%) 12m
kube-system kube-apiserver-addons-141392 250m (3%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-controller-manager-addons-141392 200m (2%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-proxy-dhft7 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system kube-scheduler-addons-141392 100m (1%) 0 (0%) 0 (0%) 0 (0%) 12m
kube-system storage-provisioner 0 (0%) 0 (0%) 0 (0%) 0 (0%) 12m
Allocated resources:
(Total limits may be over 100 percent, i.e., overcommitted.)
Resource Requests Limits
-------- -------- ------
cpu 750m (9%) 0 (0%)
memory 170Mi (0%) 170Mi (0%)
ephemeral-storage 0 (0%) 0 (0%)
hugepages-1Gi 0 (0%) 0 (0%)
hugepages-2Mi 0 (0%) 0 (0%)
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Starting 12m kube-proxy
Normal NodeHasSufficientMemory 12m (x8 over 12m) kubelet Node addons-141392 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m (x7 over 12m) kubelet Node addons-141392 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m (x7 over 12m) kubelet Node addons-141392 status is now: NodeHasSufficientPID
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal Starting 12m kubelet Starting kubelet.
Warning CgroupV1 12m kubelet Cgroup v1 support is in maintenance mode, please migrate to Cgroup v2.
Normal NodeAllocatableEnforced 12m kubelet Updated Node Allocatable limit across pods
Normal NodeHasSufficientMemory 12m kubelet Node addons-141392 status is now: NodeHasSufficientMemory
Normal NodeHasNoDiskPressure 12m kubelet Node addons-141392 status is now: NodeHasNoDiskPressure
Normal NodeHasSufficientPID 12m kubelet Node addons-141392 status is now: NodeHasSufficientPID
Normal RegisteredNode 12m node-controller Node addons-141392 event: Registered Node addons-141392 in Controller
==> dmesg <==
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff b6 0a 90 c2 1a 88 08 06
[ +0.049022] IPv4: martian source 10.244.0.1 from 10.244.0.17, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff d2 7a f4 03 56 48 08 06
[ +6.062065] IPv4: martian source 10.244.0.1 from 10.244.0.19, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 1e 8e 28 66 d4 f8 08 06
[ +0.012358] IPv4: martian source 10.244.0.1 from 10.244.0.21, on dev eth0
[ +0.000007] ll header: 00000000: ff ff ff ff ff ff d6 fd 81 c8 2b 6b 08 06
[ +0.187581] IPv4: martian source 10.244.0.1 from 10.244.0.20, on dev eth0
[ +0.000008] ll header: 00000000: ff ff ff ff ff ff fe 08 ef da e9 e7 08 06
[Sep 6 18:31] IPv4: martian source 10.244.0.1 from 10.244.0.22, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 76 9d 22 e3 d3 66 08 06
[Sep 6 18:32] IPv4: martian source 10.244.0.1 from 10.244.0.25, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 3e 86 73 e8 49 2b 08 06
[ +0.010019] IPv4: martian source 10.244.0.1 from 10.244.0.24, on dev eth0
[ +0.000005] ll header: 00000000: ff ff ff ff ff ff 6a b9 a5 7e 39 62 08 06
[ +25.833075] IPv4: martian source 10.244.0.1 from 10.244.0.26, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff ca fa 4b 24 e0 48 08 06
[ +0.000458] IPv4: martian source 10.244.0.26 from 10.244.0.2, on dev eth0
[ +0.000004] ll header: 00000000: ff ff ff ff ff ff be 94 7b dd 4b 90 08 06
[Sep 6 18:42] IPv4: martian source 10.244.0.1 from 10.244.0.34, on dev eth0
[ +0.000009] ll header: 00000000: ff ff ff ff ff ff 4e ba 34 dc ea 65 08 06
[ +23.654093] IPv4: martian source 10.244.0.37 from 10.244.0.22, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff 76 9d 22 e3 d3 66 08 06
[ +2.236229] IPv4: martian source 10.244.0.22 from 10.244.0.2, on dev eth0
[ +0.000006] ll header: 00000000: ff ff ff ff ff ff be 94 7b dd 4b 90 08 06
==> etcd [41162656997e] <==
{"level":"info","ts":"2024-09-06T18:29:51.874758Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became pre-candidate at term 1"}
{"level":"info","ts":"2024-09-06T18:29:51.874786Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgPreVoteResp from aec36adc501070cc at term 1"}
{"level":"info","ts":"2024-09-06T18:29:51.874812Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became candidate at term 2"}
{"level":"info","ts":"2024-09-06T18:29:51.874820Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc received MsgVoteResp from aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-06T18:29:51.874835Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"aec36adc501070cc became leader at term 2"}
{"level":"info","ts":"2024-09-06T18:29:51.874845Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"raft.node: aec36adc501070cc elected leader aec36adc501070cc at term 2"}
{"level":"info","ts":"2024-09-06T18:29:51.875733Z","caller":"etcdserver/server.go:2629","msg":"setting up initial cluster version using v2 API","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-06T18:29:51.876266Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-06T18:29:51.876267Z","caller":"etcdserver/server.go:2118","msg":"published local member to cluster through raft","local-member-id":"aec36adc501070cc","local-member-attributes":"{Name:addons-141392 ClientURLs:[https://192.168.49.2:2379]}","request-path":"/0/members/aec36adc501070cc/attributes","cluster-id":"fa54960ea34d58be","publish-timeout":"7s"}
{"level":"info","ts":"2024-09-06T18:29:51.876282Z","caller":"embed/serve.go:103","msg":"ready to serve client requests"}
{"level":"info","ts":"2024-09-06T18:29:51.876560Z","caller":"membership/cluster.go:584","msg":"set initial cluster version","cluster-id":"fa54960ea34d58be","local-member-id":"aec36adc501070cc","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-06T18:29:51.876620Z","caller":"etcdmain/main.go:44","msg":"notifying init daemon"}
{"level":"info","ts":"2024-09-06T18:29:51.876635Z","caller":"etcdmain/main.go:50","msg":"successfully notified init daemon"}
{"level":"info","ts":"2024-09-06T18:29:51.876638Z","caller":"api/capability.go:75","msg":"enabled capabilities for version","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-06T18:29:51.876661Z","caller":"etcdserver/server.go:2653","msg":"cluster version is updated","cluster-version":"3.5"}
{"level":"info","ts":"2024-09-06T18:29:51.877456Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-06T18:29:51.877805Z","caller":"v3rpc/health.go:61","msg":"grpc service status changed","service":"","status":"SERVING"}
{"level":"info","ts":"2024-09-06T18:29:51.878547Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"192.168.49.2:2379"}
{"level":"info","ts":"2024-09-06T18:29:51.878836Z","caller":"embed/serve.go:250","msg":"serving client traffic securely","traffic":"grpc+http","address":"127.0.0.1:2379"}
{"level":"info","ts":"2024-09-06T18:30:42.331662Z","caller":"traceutil/trace.go:171","msg":"trace[1544568972] transaction","detail":"{read_only:false; response_revision:1098; number_of_response:1; }","duration":"126.134832ms","start":"2024-09-06T18:30:42.205509Z","end":"2024-09-06T18:30:42.331644Z","steps":["trace[1544568972] 'process raft request' (duration: 111.34414ms)","trace[1544568972] 'compare' (duration: 14.607547ms)"],"step_count":2}
{"level":"info","ts":"2024-09-06T18:30:42.331954Z","caller":"traceutil/trace.go:171","msg":"trace[1301881739] transaction","detail":"{read_only:false; response_revision:1099; number_of_response:1; }","duration":"126.206049ms","start":"2024-09-06T18:30:42.205737Z","end":"2024-09-06T18:30:42.331943Z","steps":["trace[1301881739] 'process raft request' (duration: 126.09611ms)"],"step_count":1}
{"level":"info","ts":"2024-09-06T18:30:42.332075Z","caller":"traceutil/trace.go:171","msg":"trace[534260931] transaction","detail":"{read_only:false; response_revision:1100; number_of_response:1; }","duration":"125.709241ms","start":"2024-09-06T18:30:42.206351Z","end":"2024-09-06T18:30:42.332060Z","steps":["trace[534260931] 'process raft request' (duration: 125.550343ms)"],"step_count":1}
{"level":"info","ts":"2024-09-06T18:39:52.386829Z","caller":"mvcc/index.go:214","msg":"compact tree index","revision":1920}
{"level":"info","ts":"2024-09-06T18:39:52.408967Z","caller":"mvcc/kvstore_compaction.go:69","msg":"finished scheduled compaction","compact-revision":1920,"took":"21.634235ms","hash":2014098225,"current-db-size-bytes":9351168,"current-db-size":"9.4 MB","current-db-size-in-use-bytes":5054464,"current-db-size-in-use":"5.1 MB"}
{"level":"info","ts":"2024-09-06T18:39:52.409007Z","caller":"mvcc/hash.go:137","msg":"storing new hash","hash":2014098225,"revision":1920,"compact-revision":-1}
==> gcp-auth [be65c977e4e8] <==
2024/09/06 18:33:24 Ready to write response ...
2024/09/06 18:41:26 Ready to marshal response ...
2024/09/06 18:41:26 Ready to write response ...
2024/09/06 18:41:26 Ready to marshal response ...
2024/09/06 18:41:26 Ready to write response ...
2024/09/06 18:41:34 Ready to marshal response ...
2024/09/06 18:41:34 Ready to write response ...
2024/09/06 18:41:36 Ready to marshal response ...
2024/09/06 18:41:36 Ready to write response ...
2024/09/06 18:41:42 Ready to marshal response ...
2024/09/06 18:41:42 Ready to write response ...
2024/09/06 18:42:00 Ready to marshal response ...
2024/09/06 18:42:00 Ready to write response ...
2024/09/06 18:42:01 Ready to marshal response ...
2024/09/06 18:42:01 Ready to write response ...
2024/09/06 18:42:05 Ready to marshal response ...
2024/09/06 18:42:05 Ready to write response ...
2024/09/06 18:42:05 Ready to marshal response ...
2024/09/06 18:42:05 Ready to write response ...
2024/09/06 18:42:05 Ready to marshal response ...
2024/09/06 18:42:05 Ready to write response ...
2024/09/06 18:42:17 Ready to marshal response ...
2024/09/06 18:42:17 Ready to write response ...
2024/09/06 18:42:26 Ready to marshal response ...
2024/09/06 18:42:26 Ready to write response ...
==> kernel <==
18:42:38 up 25 min, 0 users, load average: 0.72, 0.42, 0.33
Linux addons-141392 5.15.0-1067-gcp #75~20.04.1-Ubuntu SMP Wed Aug 7 20:43:22 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
PRETTY_NAME="Ubuntu 22.04.4 LTS"
==> kube-apiserver [472ba18d688a] <==
W0906 18:33:16.685848 1 cacher.go:171] Terminating all watchers from cacher jobflows.flow.volcano.sh
W0906 18:33:16.686021 1 cacher.go:171] Terminating all watchers from cacher jobs.batch.volcano.sh
W0906 18:33:17.001074 1 cacher.go:171] Terminating all watchers from cacher jobtemplates.flow.volcano.sh
I0906 18:41:49.976659 1 controller.go:615] quota admission added evaluator for: volumesnapshots.snapshot.storage.k8s.io
E0906 18:41:50.363093 1 authentication.go:73] "Unable to authenticate the request" err="[invalid bearer token, serviceaccounts \"local-path-provisioner-service-account\" not found]"
I0906 18:42:05.556588 1 alloc.go:330] "allocated clusterIPs" service="headlamp/headlamp" clusterIPs={"IPv4":"10.111.111.209"}
E0906 18:42:09.457821 1 watch.go:250] "Unhandled Error" err="http2: stream closed" logger="UnhandledError"
I0906 18:42:16.616931 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0906 18:42:16.616982 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0906 18:42:16.627216 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0906 18:42:16.627270 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0906 18:42:16.638237 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0906 18:42:16.638291 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0906 18:42:16.672146 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0906 18:42:16.672188 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0906 18:42:16.681064 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1 to ResourceManager
I0906 18:42:16.681098 1 handler.go:286] Adding GroupVersion snapshot.storage.k8s.io v1beta1 to ResourceManager
I0906 18:42:17.567185 1 controller.go:615] quota admission added evaluator for: ingresses.networking.k8s.io
W0906 18:42:17.672464 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotcontents.snapshot.storage.k8s.io
W0906 18:42:17.681314 1 cacher.go:171] Terminating all watchers from cacher volumesnapshots.snapshot.storage.k8s.io
I0906 18:42:17.735935 1 alloc.go:330] "allocated clusterIPs" service="default/nginx" clusterIPs={"IPv4":"10.96.5.41"}
W0906 18:42:17.783590 1 cacher.go:171] Terminating all watchers from cacher volumesnapshotclasses.snapshot.storage.k8s.io
I0906 18:42:23.351029 1 handler.go:286] Adding GroupVersion gadget.kinvolk.io v1alpha1 to ResourceManager
W0906 18:42:24.367114 1 cacher.go:171] Terminating all watchers from cacher traces.gadget.kinvolk.io
I0906 18:42:26.198624 1 alloc.go:330] "allocated clusterIPs" service="default/hello-world-app" clusterIPs={"IPv4":"10.110.155.131"}
==> kube-controller-manager [34baa43f8805] <==
E0906 18:42:27.820470 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0906 18:42:27.881534 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="5.732811ms"
I0906 18:42:27.881614 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="default/hello-world-app-55bf9c44b4" duration="42.815µs"
W0906 18:42:27.982890 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0906 18:42:27.982926 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0906 18:42:28.517483 1 job_controller.go:568] "enqueueing job" logger="job-controller" key="ingress-nginx/ingress-nginx-admission-create" delay="0s"
I0906 18:42:28.576093 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="ingress-nginx/ingress-nginx-controller-bc57996ff" duration="9.184µs"
I0906 18:42:28.576401 1 job_controller.go:568] "enqueueing job" logger="job-controller" key="ingress-nginx/ingress-nginx-admission-patch" delay="0s"
I0906 18:42:30.721091 1 range_allocator.go:241] "Successfully synced" logger="node-ipam-controller" key="addons-141392"
I0906 18:42:31.354917 1 shared_informer.go:313] Waiting for caches to sync for resource quota
I0906 18:42:31.354948 1 shared_informer.go:320] Caches are synced for resource quota
I0906 18:42:31.577363 1 shared_informer.go:313] Waiting for caches to sync for garbage collector
I0906 18:42:31.577401 1 shared_informer.go:320] Caches are synced for garbage collector
W0906 18:42:31.606009 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0906 18:42:31.606049 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0906 18:42:33.788533 1 namespace_controller.go:187] "Namespace has been deleted" logger="namespace-controller" namespace="gadget"
W0906 18:42:34.959990 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0906 18:42:34.960025 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0906 18:42:36.818040 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0906 18:42:36.818089 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
W0906 18:42:37.233065 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0906 18:42:37.233102 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
I0906 18:42:37.307372 1 replica_set.go:679] "Finished syncing" logger="replicaset-controller" kind="ReplicaSet" key="kube-system/registry-6fb4cdfc84" duration="4.798µs"
W0906 18:42:37.667218 1 reflector.go:561] k8s.io/client-go/metadata/metadatainformer/informer.go:138: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource
E0906 18:42:37.667260 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/metadata/metadatainformer/informer.go:138: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource" logger="UnhandledError"
==> kube-proxy [7ff2536e7673] <==
I0906 18:30:04.784812 1 server_linux.go:66] "Using iptables proxy"
I0906 18:30:05.288468 1 server.go:677] "Successfully retrieved node IP(s)" IPs=["192.168.49.2"]
E0906 18:30:05.288549 1 server.go:234] "Kube-proxy configuration may be incomplete or incorrect" err="nodePortAddresses is unset; NodePort connections will be accepted on all local IPs. Consider using `--nodeport-addresses primary`"
I0906 18:30:05.683017 1 server.go:243] "kube-proxy running in dual-stack mode" primary ipFamily="IPv4"
I0906 18:30:05.683094 1 server_linux.go:169] "Using iptables Proxier"
I0906 18:30:05.690772 1 proxier.go:255] "Setting route_localnet=1 to allow node-ports on localhost; to change this either disable iptables.localhostNodePorts (--iptables-localhost-nodeports) or set nodePortAddresses (--nodeport-addresses) to filter loopback addresses" ipFamily="IPv4"
I0906 18:30:05.691173 1 server.go:483] "Version info" version="v1.31.0"
I0906 18:30:05.691197 1 server.go:485] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
I0906 18:30:05.777241 1 config.go:197] "Starting service config controller"
I0906 18:30:05.777275 1 shared_informer.go:313] Waiting for caches to sync for service config
I0906 18:30:05.777298 1 config.go:104] "Starting endpoint slice config controller"
I0906 18:30:05.777304 1 shared_informer.go:313] Waiting for caches to sync for endpoint slice config
I0906 18:30:05.777779 1 config.go:326] "Starting node config controller"
I0906 18:30:05.777796 1 shared_informer.go:313] Waiting for caches to sync for node config
I0906 18:30:05.877389 1 shared_informer.go:320] Caches are synced for endpoint slice config
I0906 18:30:05.877438 1 shared_informer.go:320] Caches are synced for service config
I0906 18:30:05.878105 1 shared_informer.go:320] Caches are synced for node config
==> kube-scheduler [bd5471ad3706] <==
E0906 18:29:54.101801 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
E0906 18:29:54.101816 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PodDisruptionBudget: failed to list *v1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User \"system:kube-scheduler\" cannot list resource \"poddisruptionbudgets\" in API group \"policy\" at the cluster scope" logger="UnhandledError"
W0906 18:29:54.101853 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope
E0906 18:29:54.101868 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Pod: failed to list *v1.Pod: pods is forbidden: User \"system:kube-scheduler\" cannot list resource \"pods\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0906 18:29:54.101917 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
W0906 18:29:54.101938 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope
E0906 18:29:54.101949 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolume: failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumes\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0906 18:29:54.101936 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0906 18:29:54.101979 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope
W0906 18:29:54.101984 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Namespace: namespaces is forbidden: User "system:kube-scheduler" cannot list resource "namespaces" in API group "" at the cluster scope
E0906 18:29:54.102004 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Namespace: failed to list *v1.Namespace: namespaces is forbidden: User \"system:kube-scheduler\" cannot list resource \"namespaces\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0906 18:29:54.102004 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicaSet: failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicasets\" in API group \"apps\" at the cluster scope" logger="UnhandledError"
W0906 18:29:54.102031 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope
W0906 18:29:54.102032 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csistoragecapacities" in API group "storage.k8s.io" at the cluster scope
W0906 18:29:54.101986 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope
E0906 18:29:54.102043 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.ReplicationController: failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User \"system:kube-scheduler\" cannot list resource \"replicationcontrollers\" in API group \"\" at the cluster scope" logger="UnhandledError"
E0906 18:29:54.102046 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIStorageCapacity: failed to list *v1.CSIStorageCapacity: csistoragecapacities.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csistoragecapacities\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
E0906 18:29:54.102054 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSINode: failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"csinodes\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
W0906 18:29:55.088882 1 reflector.go:561] runtime/asm_amd64.s:1695: failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system"
E0906 18:29:55.088925 1 reflector.go:158] "Unhandled Error" err="runtime/asm_amd64.s:1695: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"extension-apiserver-authentication\" is forbidden: User \"system:kube-scheduler\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\"" logger="UnhandledError"
W0906 18:29:55.153601 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope
E0906 18:29:55.153645 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.PersistentVolumeClaim: failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User \"system:kube-scheduler\" cannot list resource \"persistentvolumeclaims\" in API group \"\" at the cluster scope" logger="UnhandledError"
W0906 18:29:55.172477 1 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope
E0906 18:29:55.172509 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.StorageClass: failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User \"system:kube-scheduler\" cannot list resource \"storageclasses\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError"
I0906 18:29:56.899717 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file
==> kubelet <==
Sep 06 18:42:32 addons-141392 kubelet[2432]: I0906 18:42:32.109050 2432 reconciler_common.go:288] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/26ce4a9c-23cf-4be3-ab7b-37124885de90-webhook-cert\") on node \"addons-141392\" DevicePath \"\""
Sep 06 18:42:32 addons-141392 kubelet[2432]: I0906 18:42:32.116811 2432 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26ce4a9c-23cf-4be3-ab7b-37124885de90" path="/var/lib/kubelet/pods/26ce4a9c-23cf-4be3-ab7b-37124885de90/volumes"
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.039357 2432 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbnzv\" (UniqueName: \"kubernetes.io/projected/72b54610-0af5-45ef-80de-28ee45b90ba8-kube-api-access-zbnzv\") pod \"72b54610-0af5-45ef-80de-28ee45b90ba8\" (UID: \"72b54610-0af5-45ef-80de-28ee45b90ba8\") "
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.039405 2432 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/72b54610-0af5-45ef-80de-28ee45b90ba8-gcp-creds\") pod \"72b54610-0af5-45ef-80de-28ee45b90ba8\" (UID: \"72b54610-0af5-45ef-80de-28ee45b90ba8\") "
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.039489 2432 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54610-0af5-45ef-80de-28ee45b90ba8-gcp-creds" (OuterVolumeSpecName: "gcp-creds") pod "72b54610-0af5-45ef-80de-28ee45b90ba8" (UID: "72b54610-0af5-45ef-80de-28ee45b90ba8"). InnerVolumeSpecName "gcp-creds". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.041035 2432 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72b54610-0af5-45ef-80de-28ee45b90ba8-kube-api-access-zbnzv" (OuterVolumeSpecName: "kube-api-access-zbnzv") pod "72b54610-0af5-45ef-80de-28ee45b90ba8" (UID: "72b54610-0af5-45ef-80de-28ee45b90ba8"). InnerVolumeSpecName "kube-api-access-zbnzv". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.139764 2432 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-zbnzv\" (UniqueName: \"kubernetes.io/projected/72b54610-0af5-45ef-80de-28ee45b90ba8-kube-api-access-zbnzv\") on node \"addons-141392\" DevicePath \"\""
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.139808 2432 reconciler_common.go:288] "Volume detached for volume \"gcp-creds\" (UniqueName: \"kubernetes.io/host-path/72b54610-0af5-45ef-80de-28ee45b90ba8-gcp-creds\") on node \"addons-141392\" DevicePath \"\""
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.673959 2432 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx7n8\" (UniqueName: \"kubernetes.io/projected/03e8d3f6-b271-4c1d-a933-77656192d8eb-kube-api-access-jx7n8\") pod \"03e8d3f6-b271-4c1d-a933-77656192d8eb\" (UID: \"03e8d3f6-b271-4c1d-a933-77656192d8eb\") "
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.674018 2432 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kzh6\" (UniqueName: \"kubernetes.io/projected/23346424-7dc9-4199-af57-8fffc4bcfe51-kube-api-access-4kzh6\") pod \"23346424-7dc9-4199-af57-8fffc4bcfe51\" (UID: \"23346424-7dc9-4199-af57-8fffc4bcfe51\") "
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.675839 2432 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03e8d3f6-b271-4c1d-a933-77656192d8eb-kube-api-access-jx7n8" (OuterVolumeSpecName: "kube-api-access-jx7n8") pod "03e8d3f6-b271-4c1d-a933-77656192d8eb" (UID: "03e8d3f6-b271-4c1d-a933-77656192d8eb"). InnerVolumeSpecName "kube-api-access-jx7n8". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.675885 2432 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23346424-7dc9-4199-af57-8fffc4bcfe51-kube-api-access-4kzh6" (OuterVolumeSpecName: "kube-api-access-4kzh6") pod "23346424-7dc9-4199-af57-8fffc4bcfe51" (UID: "23346424-7dc9-4199-af57-8fffc4bcfe51"). InnerVolumeSpecName "kube-api-access-4kzh6". PluginName "kubernetes.io/projected", VolumeGidValue ""
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.775143 2432 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-jx7n8\" (UniqueName: \"kubernetes.io/projected/03e8d3f6-b271-4c1d-a933-77656192d8eb-kube-api-access-jx7n8\") on node \"addons-141392\" DevicePath \"\""
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.775185 2432 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-4kzh6\" (UniqueName: \"kubernetes.io/projected/23346424-7dc9-4199-af57-8fffc4bcfe51-kube-api-access-4kzh6\") on node \"addons-141392\" DevicePath \"\""
Sep 06 18:42:37 addons-141392 kubelet[2432]: I0906 18:42:37.987386 2432 scope.go:117] "RemoveContainer" containerID="e05cbbcfa65246fc4b39126a3ba491189b4aefc7b0237ef99f275c231fa08cab"
Sep 06 18:42:38 addons-141392 kubelet[2432]: I0906 18:42:38.003462 2432 scope.go:117] "RemoveContainer" containerID="e05cbbcfa65246fc4b39126a3ba491189b4aefc7b0237ef99f275c231fa08cab"
Sep 06 18:42:38 addons-141392 kubelet[2432]: E0906 18:42:38.004451 2432 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: e05cbbcfa65246fc4b39126a3ba491189b4aefc7b0237ef99f275c231fa08cab" containerID="e05cbbcfa65246fc4b39126a3ba491189b4aefc7b0237ef99f275c231fa08cab"
Sep 06 18:42:38 addons-141392 kubelet[2432]: I0906 18:42:38.004501 2432 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"e05cbbcfa65246fc4b39126a3ba491189b4aefc7b0237ef99f275c231fa08cab"} err="failed to get container status \"e05cbbcfa65246fc4b39126a3ba491189b4aefc7b0237ef99f275c231fa08cab\": rpc error: code = Unknown desc = Error response from daemon: No such container: e05cbbcfa65246fc4b39126a3ba491189b4aefc7b0237ef99f275c231fa08cab"
Sep 06 18:42:38 addons-141392 kubelet[2432]: I0906 18:42:38.004526 2432 scope.go:117] "RemoveContainer" containerID="8ea787065b670dd7cdb148ecb0bd08e1cb78613ad289e66cff17497e4b649a32"
Sep 06 18:42:38 addons-141392 kubelet[2432]: I0906 18:42:38.016179 2432 scope.go:117] "RemoveContainer" containerID="8ea787065b670dd7cdb148ecb0bd08e1cb78613ad289e66cff17497e4b649a32"
Sep 06 18:42:38 addons-141392 kubelet[2432]: E0906 18:42:38.016739 2432 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = Unknown desc = Error response from daemon: No such container: 8ea787065b670dd7cdb148ecb0bd08e1cb78613ad289e66cff17497e4b649a32" containerID="8ea787065b670dd7cdb148ecb0bd08e1cb78613ad289e66cff17497e4b649a32"
Sep 06 18:42:38 addons-141392 kubelet[2432]: I0906 18:42:38.016779 2432 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"docker","ID":"8ea787065b670dd7cdb148ecb0bd08e1cb78613ad289e66cff17497e4b649a32"} err="failed to get container status \"8ea787065b670dd7cdb148ecb0bd08e1cb78613ad289e66cff17497e4b649a32\": rpc error: code = Unknown desc = Error response from daemon: No such container: 8ea787065b670dd7cdb148ecb0bd08e1cb78613ad289e66cff17497e4b649a32"
Sep 06 18:42:38 addons-141392 kubelet[2432]: I0906 18:42:38.116864 2432 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03e8d3f6-b271-4c1d-a933-77656192d8eb" path="/var/lib/kubelet/pods/03e8d3f6-b271-4c1d-a933-77656192d8eb/volumes"
Sep 06 18:42:38 addons-141392 kubelet[2432]: I0906 18:42:38.117341 2432 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23346424-7dc9-4199-af57-8fffc4bcfe51" path="/var/lib/kubelet/pods/23346424-7dc9-4199-af57-8fffc4bcfe51/volumes"
Sep 06 18:42:38 addons-141392 kubelet[2432]: I0906 18:42:38.117698 2432 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72b54610-0af5-45ef-80de-28ee45b90ba8" path="/var/lib/kubelet/pods/72b54610-0af5-45ef-80de-28ee45b90ba8/volumes"
==> storage-provisioner [f1925f13e831] <==
I0906 18:30:09.778885 1 storage_provisioner.go:116] Initializing the minikube storage provisioner...
I0906 18:30:09.793749 1 storage_provisioner.go:141] Storage provisioner initialized, now starting service!
I0906 18:30:09.793787 1 leaderelection.go:243] attempting to acquire leader lease kube-system/k8s.io-minikube-hostpath...
I0906 18:30:09.878754 1 leaderelection.go:253] successfully acquired lease kube-system/k8s.io-minikube-hostpath
I0906 18:30:09.878943 1 controller.go:835] Starting provisioner controller k8s.io/minikube-hostpath_addons-141392_ab13a48b-83d0-48f0-9363-b9ffc85eaecb!
I0906 18:30:09.880715 1 event.go:282] Event(v1.ObjectReference{Kind:"Endpoints", Namespace:"kube-system", Name:"k8s.io-minikube-hostpath", UID:"314649fa-59f5-489a-9e6a-f88ab7492d72", APIVersion:"v1", ResourceVersion:"636", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' addons-141392_ab13a48b-83d0-48f0-9363-b9ffc85eaecb became leader
I0906 18:30:09.979756 1 controller.go:884] Started provisioner controller k8s.io/minikube-hostpath_addons-141392_ab13a48b-83d0-48f0-9363-b9ffc85eaecb!
-- /stdout --
helpers_test.go:254: (dbg) Run: out/minikube-linux-amd64 status --format={{.APIServer}} -p addons-141392 -n addons-141392
helpers_test.go:261: (dbg) Run: kubectl --context addons-141392 get po -o=jsonpath={.items[*].metadata.name} -A --field-selector=status.phase!=Running
helpers_test.go:272: non-running pods: busybox
helpers_test.go:274: ======> post-mortem[TestAddons/parallel/Registry]: describe non-running pods <======
helpers_test.go:277: (dbg) Run: kubectl --context addons-141392 describe pod busybox
helpers_test.go:282: (dbg) kubectl --context addons-141392 describe pod busybox:
-- stdout --
Name: busybox
Namespace: default
Priority: 0
Service Account: default
Node: addons-141392/192.168.49.2
Start Time: Fri, 06 Sep 2024 18:33:24 +0000
Labels: integration-test=busybox
Annotations: <none>
Status: Pending
IP: 10.244.0.28
IPs:
IP: 10.244.0.28
Containers:
busybox:
Container ID:
Image: gcr.io/k8s-minikube/busybox:1.28.4-glibc
Image ID:
Port: <none>
Host Port: <none>
Command:
sleep
3600
State: Waiting
Reason: ImagePullBackOff
Ready: False
Restart Count: 0
Environment:
GOOGLE_APPLICATION_CREDENTIALS: /google-app-creds.json
PROJECT_ID: this_is_fake
GCP_PROJECT: this_is_fake
GCLOUD_PROJECT: this_is_fake
GOOGLE_CLOUD_PROJECT: this_is_fake
CLOUDSDK_CORE_PROJECT: this_is_fake
Mounts:
/google-app-creds.json from gcp-creds (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-b5tfk (ro)
Conditions:
Type Status
PodReadyToStartContainers True
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
kube-api-access-b5tfk:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
gcp-creds:
Type: HostPath (bare host directory volume)
Path: /var/lib/minikube/google_application_credentials.json
HostPathType: File
QoS Class: BestEffort
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 9m14s default-scheduler Successfully assigned default/busybox to addons-141392
Normal Pulling 7m52s (x4 over 9m13s) kubelet Pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
Warning Failed 7m52s (x4 over 9m13s) kubelet Failed to pull image "gcr.io/k8s-minikube/busybox:1.28.4-glibc": Error response from daemon: Head "https://gcr.io/v2/k8s-minikube/busybox/manifests/1.28.4-glibc": unauthorized: authentication failed
Warning Failed 7m52s (x4 over 9m13s) kubelet Error: ErrImagePull
Warning Failed 7m29s (x6 over 9m13s) kubelet Error: ImagePullBackOff
Normal BackOff 4m10s (x21 over 9m13s) kubelet Back-off pulling image "gcr.io/k8s-minikube/busybox:1.28.4-glibc"
-- /stdout --
helpers_test.go:285: <<< TestAddons/parallel/Registry FAILED: end of post-mortem logs <<<
helpers_test.go:286: ---------------------/post-mortem---------------------------------
--- FAIL: TestAddons/parallel/Registry (72.35s)